Abstract
Classification is a quite relevant task within data mining area. This task is not trivial and some difficulties can arise depending on the nature of the problem. Multiple classifier systems have been used to construct ensembles of base classifiers in order to solve or alleviate some of those problems. One of the most current problems that is being studied in recent years is how to learn when the datasets are too large or when new information can arrive at any time. In that case, incremental learning is an approach that can be used. Some works have used multiple classifier system to learn in an incremental way and the results are very promising. The aim of this paper is to propose a method for improving the classification (or prediction) accuracy reached by multiple classifier systems in this context.
This work has been partially supported by the FPI program and the MOISES-TA project, number TIN2005-08832-C03, of the MEC, Spain.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Breiman, L.: Bagging predictors. Machine Learning 24(2), 123–140 (1996)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proc. of the 13th Int. Conf. on Machine Learning, pp. 146–148 (1996)
Wang, H., Fan, W., Yu, P.S., Han, J.: Mining concept-drifting data streams using ensemble classifiers. In: Proc. 9th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 226–235. ACM Press, New York (2003)
Breiman, L.: Pasting small votes for classification in large databases and on-line. Machine Learning 36, 85–103 (1999)
Street, W.N., Kim, Y.: A streaming ensemble algorithm (SEA) for large-scale classification. In: Proc. of the Seventh ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 377–382. ACM Press, New York (2001)
Oza, N.C.: Online bagging and boosting. In: Proc. of the IEEE Int. Conf. on Systems, Man and Cybernetics, pp. 2340–2345. IEEE Press, Los Alamitos (2005)
Hu, Q., Yu, D., Wang, M.: Constructing rough decision forests. In: Ślęzak, D., Yao, J., Peters, J.F., Ziarko, W., Hu, X. (eds.) RSFDGrC 2005. LNCS (LNAI), vol. 3642, pp. 147–156. Springer, Heidelberg (2005)
Hu, X.: Ensembles of classifiers based on rough sets theory and set-oriented database operations. In: Proc. of the IEEE In. Conf. on Granular Computing, pp. 67–73. IEEE Press, Los Alamitos (2006)
Polikar, R., Udpa, L., Udpa, S., Honavar, V.: Learn++: An incremental learning algorithm for supervised neural networks. IEEE Transactions on Systems, Man, and Cybernetics 31, 497–508 (2001)
Fern, A., Givan, R.: Online ensemble learning: An empirical study. Machine Learning 53, 71–109 (2003)
Fan, W., Stolfo, S.J., Zhang, J.: The application of adaboost for distributed, scalable and on-line learning. In: Proc. of the fifth ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 362–366. ACM Press, New York (1999)
Gama, J., Fernandes, R., Rocha, R.: Decision trees for mining data streams. Intelligent Data Analysis 10(1), 23–45 (2006)
Ramos-Jiménez, G., Campo-Ávila, J., Morales-Bueno, R.: Induction of decision trees using an internal control of induction. In: Cabestany, J., Prieto, A.G., Sandoval, F. (eds.) IWANN 2005. LNCS, vol. 3512, pp. 795–803. Springer, Heidelberg (2005)
Ramos-Jiménez, G., del Campo-Ávila, J., Morales-Bueno, R.: FE-CIDIM: Fast ensemble of CIDIM classifiers. International Journal of Systems Science 37, 939–947 (2006)
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Francisco (1993)
Ortega, J.: Exploiting multiple existing models and learning algorithms. In: Proc. of the Integrating Multiple Learned Models for Improving and Scaling Machine Learning Algorithms Workshop (1996)
Domingos, P., Hulten, G.: Mining high-speed data streams. In: Proc. of the 6th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining, pp. 71–80. ACM Press, New York (2000)
Ramos-Jiménez, G., del Campo-Ávila, J., Morales-Bueno, R.: Incremental algorithm driven by error margins. In: Todorovski, L., Lavrač, N., Jantke, K.P. (eds.) DS 2006. LNCS (LNAI), vol. 4265, pp. 358–362. Springer, Heidelberg (2006)
del Campo-Ávila, J., Ramos-Jiménez, G., Gama, J., Morales-Bueno, R.: Improving the performance of an incremental algorithm driven by error margins. In: Intelligent Data Analysis (2007) (to appear)
Blake, C., Merz, C.J.: UCI repository of machine learning databases. University of California, Department of Information and Computer Science (2000)
Demšar, J.: Statistical comparisons of classifiers over multiple datasets. Journal of Machine Learning Research 7, 1–30 (2006)
Witten, I.H., Frank, E.: Data Mining: Practical machine learning tools with Java implementations. Morgan Kaufmann, San Francisco (2000)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
del Campo-Ávila, J., Ramos-Jiménez, G., Morales-Bueno, R. (2007). Incremental Learning with Multiple Classifier Systems Using Correction Filters for Classification. In: R. Berthold, M., Shawe-Taylor, J., Lavrač, N. (eds) Advances in Intelligent Data Analysis VII. IDA 2007. Lecture Notes in Computer Science, vol 4723. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74825-0_10
Download citation
DOI: https://doi.org/10.1007/978-3-540-74825-0_10
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-74824-3
Online ISBN: 978-3-540-74825-0
eBook Packages: Computer ScienceComputer Science (R0)