Abstract
Existing classification rule learning algorithms use mainly greedy heuristic search to find regularities in datasets for classification. In recent years, extensive research on association rule mining was performed in the machine learning community on learning rules by using exhaustive search. The main objective is to find all rules in data that satisfy the user-specified minimum support and minimum confidence constraints. Although the whole set of rules may not be used directly for accurate classification, effective and efficient classifiers have been built using these, so called, classification association rules.
In this paper, we compare “classical” classification rule learning algorithms that use greedy heuristic search to produce the final classifier with a class association rule learner that uses constrained exhaustive search to find classification rules on “well known” datasets. We propose a simple method to extract class association rules by simple pruning to form an accurate classifier. This is a preliminary study that aims to show that an adequate choice of the “right” class association rules by considering the dependent (class) attribute distribution of values can produce a compact, understandable and relatively accurate classifier. We have performed experiments on 12 datasets from UCI Machine Learning Database Repository and compared the results with well-known rule-based and tree-based classification algorithms. Experimental results show that our method was consistent and comparative with other well-known classification algorithms. Although not achieving the best results in terms of classification accuracy, our method is relatively simple and produces compact and understandable classifiers by exhaustively searching the entire example space.
Keywords
- Attribute
- Frequent itemset
- Minimum support
- Minimum confidence
- Class association rules (CAR)
- Associative classification
This is a preview of subscription content, access via your institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Agrawal, R., Srikant, R.: Fast algorithms for mining association rules. In: VLDB 1994 Proceedings of the 20th International Conference on Very Large Data Bases, Chile, pp. 487–499 (1994)
Ali, K., Manganaris, S., Srikant, R.: Partial classification using association rules. In: Proceedings of KDD-1997, U.S.A., pp. 115–118 (1997)
Baralis, E., Cagliero, L., Garza, P.: A novel pattern-based Bayesian classifier. IEEE Trans. Knowl. Data Eng. 25(12), 2780–2795 (2013)
Bayardo, R.J.: Brute-force mining of high-confidence classification rules. In: Proceedings of the Third International Conference on Knowledge Discovery and Data Mining, U.S.A., pp. 123–126 (1997)
Breiman, L.: Random Forests. Mach. Learn. 45(1), 5–32 (2001)
Cendrowska, J.: PRISM: an algorithm for inducing modular rules. Int. J. Man-Mach. Stud. 27(4), 349–370 (1987)
Chen, G., Liu, H., Yu, L., Wei, Q., Zhang, X.: A new approach to classification based on association rule mining. Decis. Support Syst. 42(2), 674–689 (2006)
Clark, P., Niblett, T.: The CN2 induction algorithm. Mach. Learn. 3(4), 261–283 (1989)
Cohen, W.W.: Fast effective rule induction. In: ICML 1995 Proceedings of the Twelfth International Conference on Machine Learning, Tahoe City, California, pp. 115–123 (1995)
Dua, D., Graff, C.: UCI Machine Learning Repository. University of California, Irvine (2019)
Frank, E., Witten, I.: Generating accurate rule sets without global optimization. In: Fifteenth International Conference on Machine Learning, USA, pp. 144–151 (1998)
Holte, R.: Very simple classification rules perform well on most commonly used datasets. Mach. Learn. 11(1), 63–91 (1993)
Kohavi, R.: The power of decision tables. In: 8th European Conference on Machine Learning, Heraclion, Crete, Greece, pp. 174–189 (1995)
Lent, B., Swami, A., Widom, J.: Clustering association rules. In: ICDE 1997 Proceedings of the Thirteenth International Conference on Data Engineering, England, pp. 220–231 (1997)
Li, W., Han, J., Pei, J.: CMAR: accurate and efficient classification based on multiple class-association rules. In: Proceedings of the 1st IEEE International Conference on Data Mining (ICDM 2001), San Jose, California, USA, pp. 369–376 (2001)
Liu, B., Hsu, W., Ma, Y.: Integrating classification and association rule mining. In: Proceedings of the 4th International Conference on Knowledge Discovery and Data Mining (KDD 1998), New York, USA, pp. 80–86 (1998)
Quinlan, J.: C4.5: Programs for machine learning. Mach. Learn. 16(3), 235–240 (1993)
Xiaoxin, Y., Jiawei, H. CPAR: classification based on predictive association rules. In: Proceedings of the SIAM International Conference on Data Mining, San Francisco, U.S.A., pp. 331–335 (2003)
Zhang, M., Zhou Z.: A k-nearest neighbor based algorithm for multi-label classification. In: Proceedings of the 1st IEEE International Conference on Granular Computing (GrC 2005), Beijing, China, vol. 2, pp. 718–721 (2005)
Zhou, Z., Liu, X.: Training cost-sensitive neural networks with methods addressing the class imbalance problem. IEEE Trans. Knowl. Data Eng. 18(1), 63–77 (2006)
Acknowledgement
The authors gratefully acknowledge the European Commission for funding the InnoRenew CoE project (Grant Agreement #739574) under the Horizon2020 Widespread-Teaming program and and the Republic of Slovenia (Investment funding of the Republic of Slovenia and the European Union of the European Regional Development Fund).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Mattiev, J., Kavšek, B. (2019). Simple and Accurate Classification Method Based on Class Association Rules Performs Well on Well-Known Datasets. In: Nicosia, G., Pardalos, P., Umeton, R., Giuffrida, G., Sciacca, V. (eds) Machine Learning, Optimization, and Data Science. LOD 2019. Lecture Notes in Computer Science(), vol 11943. Springer, Cham. https://doi.org/10.1007/978-3-030-37599-7_17
Download citation
DOI: https://doi.org/10.1007/978-3-030-37599-7_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-37598-0
Online ISBN: 978-3-030-37599-7
eBook Packages: Computer ScienceComputer Science (R0)