Abstract
The combination of classifier has long been proposed as a method to improve the accuracy achieved in isolation by a single classifier. Most of the extant works focus on how to generate a group of “good” base classifiers, such as AdaBoost and Bagging. We are interested in the method of combining multiple classifiers. In contrast to such popular used method as vote, we regard the classifier combination problem as a classification problem. From the perspective of pattern recognition, the base classifiers can also be regarded as a feature extraction method. In theory, any of classifiers can be used to combine the base classifiers as long as they are able to treat the outputs of base classifiers. More generally, the combination model is also able to deal with other machine learning problems including cluster and regression task, that is Learner Combination via Learner(LCL) model. A large empirical study shows that, comparing with majority vote(Bagging) and weighted majority vote(AdaBoost), CCC can significantly improve the performance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Hansen, L.K.: Neural Network Ensembles. IEE Transactions on Pattern Analysis and Machine Intelligence 12, 993–1001 (1990)
Mao, J., Mohiuddin, K.M.: Improving OCR Performance Using Character Degradation Models and Boosting Algorithm. Pattern Recognition Letters 18, 1415–1419 (1997)
Chen, K., Wang, L., Chi, H.S.: Methods of Combining Multiple Classifiers with Different Features and Their Applications to Text-independent Speaker Identification. International Journal of Pattern Recognition and Artificial Intelligence 11(3), 417–445 (1997)
Shimshoni, Y., Intrator, N.: Classification of Seismic Signals by Integrating Ensembles of Neural Networks. IEEE Transactions on Signal Processing 46(5), 1194–1201 (1998)
Breiman, L.: Bagging Predictors. Machine Learning 24(2), 123–140 (1996)
Freund, Y.: Boosting a Weak Learning Algorithm by Majority. Information and Computation 121(2), 256–285 (1995)
Freund, Y., Schapire, R.E.: A Decision-theoretic Generalization of On-line Learning and an Application to Boosting. Journal of Computer and System Sciences 55(1), 119–139 (1997)
Ho, T.K.: The Random Subspace Method for Constructing Decision Forests. IEEE Transactions on Pattern Analysis and Machine Intelligence 20(8), 832–844 (1998)
Breiman, L.: Random forests. Machine Learning 45(1), 5–32 (2001)
Rodriguez, J.J., Kuncheva, L.I.: Rotation Forest: A New Classifier Ensemble Method. IEEE Transactions on Pattern Analysis and Machine Intelligence 28(10), 1619–1630 (2006)
Fumera, G., Roli, F.: A Theoretical and Experimental Analysis of Linear Combiners for Multiple Classifier Systems. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(6), 942–956 (2005)
Huang, Y.S., Suen, C.Y.: Method of Combining Multiple Experts for the Recognition of Unconstrained Handwritten Numerals. IEEE Transactions on Pattern Analysis and Machine Intelligence 17(1), 90–94 (1995)
Wernecke, K.D.: A Coupling Procedure for the Discrimination of Mixed Data. Biometrics 48(2), 497–506 (1992)
Merz, C.J.: Using Correspondence Analysis to Combine Classifiers. Machine Learning 36(1-2), 33–58 (1999)
Kuncheva, L.I.: Combining Pattern Classifiers: Methods and Algorithms. Wiley, Hoboken (2004)
Duda, R.O., Hart, P.E., Stork, D.G.: Pattern Classification, 2nd edn. Wiley, Chichester (2000)
Zhou, Z.H., Wu, J.X., Tang, W.: Ensembling Neural Networks: Many Could Be Better Than All. Artificial Intelligence 137(1-2), 239–263 (2002)
Zhou, Z.H., Jiang, Y., Chen, S.F.: Extracting Symbolic Rules from Trained Neural Network Ensembles. Ai Communications 16(1), 3–15 (2003)
Witten, I., Frank, E.: Data Mining: Practical Machine Learning Tools and Techniques. Morgan Kaufmann, San Francisco (2005)
Blake, C.L., Merz, C.J.: UCI Repository of Machine Learning Databases (1998), http://www.ics.uci.edu/mlearn/MLRepository.html
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2011 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Lu, CY. (2011). CCC: Classifier Combination via Classifier. In: Huang, DS., Gan, Y., Bevilacqua, V., Figueroa, J.C. (eds) Advanced Intelligent Computing. ICIC 2011. Lecture Notes in Computer Science, vol 6838. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-24728-6_14
Download citation
DOI: https://doi.org/10.1007/978-3-642-24728-6_14
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-24727-9
Online ISBN: 978-3-642-24728-6
eBook Packages: Computer ScienceComputer Science (R0)