Abstract
This paper presents six filtration algorithms for the pruning of the unordered sets of regression rules. Three of these algorithms aim at the elimination of the rules which cover similar subsets of examples, whereas the other three ones aim at the optimization of the rule sets according to the prediction accuracy. The effectiveness of the filtration algorithms was empirically tested for 5 different rule learning heuristics on 35 benchmark datasets. The results show that, depending on the filtration algorithm, the reduction of the number of rules fluctuates on average between 10% and 50% and in most cases it does not cause statistically significant degradation in the accuracy of predictions.
Keywords
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Fürnkranz, J.: Pruning algorithms for rule learning. Machine Learning 27(2), 139–172 (1997)
Bramer, M.: Avoiding overfitting of decision trees. In: Principles of Data Mining, pp. 119–134. Springer, London (2007)
Bruha, I.: From machine learning to knowledge discovery: Survey of preprocessing and postprocessing. Intelligent Data Analysis 4(3,4), 363–374 (2000)
Sikora, M.: Rule Quality Measures in Creation and Reduction of Data Rule Models. In: Greco, S., Hata, Y., Hirano, S., Inuiguchi, M., Miyamoto, S., Nguyen, H.S., Słowiński, R. (eds.) RSCTC 2006. LNCS (LNAI), vol. 4259, pp. 716–725. Springer, Heidelberg (2006)
Sikora, M.: Decision Rule-Based Data Models Using TRS and NetTRS – Methods and Algorithms. In: Peters, J.F., Skowron, A. (eds.) Transactions on Rough Sets XI. LNCS, vol. 5946, pp. 130–160. Springer, Heidelberg (2010)
Sikora, M., Wróbel, Ł.: Data-driven adaptive selection of rule quality measures for improving rule induction and filtration algorithms. International Journal of General Systems 42(4) (2013) (to appear)
Ågotnes, T., Komorowski, J., Løken, T.: Taming Large Rule Models in Rough Set Approaches. In: Żytkow, J.M., Rauch, J. (eds.) PKDD 1999. LNCS (LNAI), vol. 1704, pp. 193–203. Springer, Heidelberg (1999)
Dembczyński, K., Kotłowski, W., Słowiński, R.: Solving Regression by Learning an Ensemble of Decision Rules. In: Rutkowski, L., Tadeusiewicz, R., Zadeh, L.A., Zurada, J.M. (eds.) ICAISC 2008. LNCS (LNAI), vol. 5097, pp. 533–544. Springer, Heidelberg (2008)
Friedman, J., Popescu, B.: Predictive learning via rule ensembles. The Annals of Applied Statistics, 916–954 (2008)
Janssen, F., Fürnkranz, J.: Heuristic rule-based regression via dynamic reduction to classification. In: Walsh, T. (ed.) Proceedings of the 22nd International Joint Conference on Artificial Intelligence (IJCAI 2011), pp. 1330–1335 (2011)
Sikora, M., Skowron, A., Wróbel, Ł.: Rule Quality Measure-Based Induction of Unordered Sets of Regression Rules. In: Ramsay, A., Agre, G. (eds.) AIMSA 2012. LNCS, vol. 7557, pp. 162–171. Springer, Heidelberg (2012)
Ishibuchi, H., Yamamoto, T.: Effects of three-objective genetic rule selection on the generalization ability of fuzzy rule-based systems. LNCS, pp. 608–622 (2003)
Andersen, T., Martinez, T.: NP-completeness of minimum rule sets. In: Proceedings of the 10th International Symposium on Computer and Information Sciences, pp. 411–418 (1995)
Øhrn, A., Ohno-Machado, L., Rowland, T.: Building manageable rough set classifiers. In: Proceedings of the AMIA Symposium, American Medical Informatics Association, p. 543 (1998)
Gamberger, D., Lavrač, N.: Confirmation Rule Sets. In: Zighed, D.A., Komorowski, J., Žytkow, J. (eds.) PKDD 2000. LNCS (LNAI), vol. 1910, pp. 34–43. Springer, Heidelberg (2000)
Strehl, A., Gupta, G., Ghosh, J.: Distance based clustering of association rules. Proceedings ANNIE 1999 9, 759–764 (1999)
Tsumoto, S., Hirano, S.: Visualization of rule’s similarity using multidimensional scaling. In: Third IEEE International Conference on Data Mining, ICDM 2003, pp. 339–346. IEEE (2003)
Sikora, M., Gruca, A.: Induction and selection of the most interesting gene ontology based multiattribute rules for descriptions of gene groups. Pattern Recognition Letters 32(2), 258–269 (2011)
Bayardo, R.J., Agrawal, R.: Mining the most interesting rules. In: KDD 1999: Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 145–154. ACM Press, New York (1999)
Brzezińska, I., Greco, S., Słowiński, R.: Mining pareto-optimal rules with respect to support and confirmation or support and anti-support. Engineering Applications of Artificial Intelligence 20(5), 587–600 (2007)
Demšar, J.: Statistical comparisons of classifiers over multiple data sets. J. Mach. Learn. Res. 7, 1–30 (2006)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Wróbel, Ł., Sikora, M., Skowron, A. (2012). Algorithms for Filtration of Unordered Sets of Regression Rules. In: Sombattheera, C., Loi, N.K., Wankar, R., Quan, T. (eds) Multi-disciplinary Trends in Artificial Intelligence. MIWAI 2012. Lecture Notes in Computer Science(), vol 7694. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-35455-7_26
Download citation
DOI: https://doi.org/10.1007/978-3-642-35455-7_26
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-35454-0
Online ISBN: 978-3-642-35455-7
eBook Packages: Computer ScienceComputer Science (R0)