Abstract
When a multiple classifier system is employed, one of the most popular methods to accomplish the classifier fusion is the simple majority voting. However, when the performance of the ensemble members is not uniform, the efficiency of this type of voting generally results affected negatively. In this paper, new functions for dynamic weighting in classifier fusion are introduced. Experimental results demonstrate the advantages of these novel strategies over the simple voting scheme.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Bauer, E., Kohavi, R.: An empirical comparison of voting classification algorithms: bagging, boosting, and variants. Machine Learning 36, 105–139 (1999)
Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)
Breiman, L.: Arcing classifiers. Annals of Statistics 26, 801–823 (1998)
Dasarathy, B.V.: Nearest Neighbor Norms: NN Pattern Classification Techniques. IEEE Computer Society Press, Los Alamos (1991)
Demšar, J.: Statistical comparison of classifiers over multiple data sets. Journal of Machine Learning Research 7, 1–30 (2006)
Dudani, S.A.: The distance weighted k-nearest neighbor rule. IEEE Trans. on System, Man, and Cybernetics 6, 325–327 (1976)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proc. 13th Intl. Conf. on Machine Learning, pp. 148–156 (1996)
Friedman, J., Hastie, T., Tibshirani, R.: Additive logistic regression: a statistical view of boosting, Technical report, Stanford University (1998)
Hansen, L.K., Salomon, P.: Neural network ensembles. IEEE Trans. on Pattern Analysis and Machine Intelligence 12, 993–1001 (1990)
Kuncheva, L.I., Bezdek, J.C., Duin, R.P.W.: Decision templates for multiple classifier fusion. Pattern Recognition 34, 299–314 (2001)
Matan, O.: On voting ensembles of classifiers. In: Proc. 13th Natl. Conf. on Artificial Intelligence, pp. 84–88 (1996)
Raudys, S.: Trainable fusion rules. I. Large sample size case. Neural Networks 19, 1506–1516 (2006)
Raudys, S.: Trainable fusion rules. II. Small sample-size effects. Neural Networks 19, 1517–1527 (2006)
Shepard, R.N.: Toward a universal law of generalization for psychological science. Science 237, 1317–1323 (1987)
Woods, K., Kegelmeyer Jr., W.P., Bowyer, K.: Combination of multiple classifiers using local accuracy estimates. IEEE Trans. on Pattern Analysis and Machine Intelligence 19, 405–410 (1997)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Valdovinos, R.M., Sánchez, J.S. (2009). Combining Multiple Classifiers with Dynamic Weighted Voting. In: Corchado, E., Wu, X., Oja, E., Herrero, Á., Baruque, B. (eds) Hybrid Artificial Intelligence Systems. HAIS 2009. Lecture Notes in Computer Science(), vol 5572. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02319-4_61
Download citation
DOI: https://doi.org/10.1007/978-3-642-02319-4_61
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02318-7
Online ISBN: 978-3-642-02319-4
eBook Packages: Computer ScienceComputer Science (R0)