Abstract
In this paper, the neural network ensemble algorithm is proposed to solve the problem of the mislabeled data in the tri-training process. Firstly, we analyze the advantage of the neural network ensemble, and then introduce it to correct the mislabeled data to improve the quality of the enlarged training set, so the precision and generalization of learns is improved. Experimental results on UCI data sets indicate that the classification performance of the proposed algorithm is 22.87% higher than that of the tri-training algorithm under the four kinds of the unlabeled rates. The proposed algorithm could effectively exploit unlabeled data to enhance the learning performance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Miao, Z.M., Zhao, L.W., Hu, G.Y., Wang, Q.: Semi-supervised Learning Based on One-class Classification. Pattern Recognition and Artificial Intelligence 22(6), 924–930 (2009)
Dempster, A.P., Laird, N.M., Rubin, D.B.: Maximum likelihood from incomplete data via the EM algorithm. J. Royal Statistical Soc., Series B 39(1), 1–38 (1977)
Joachims, T.: Transductive inference for text classification using support vector machines. In: Proc. 16th Int’l Conf. Machine Learning, pp. 200–209 (1999)
Blum, A., Chawla, S.: Learning from labeled and unlabeled data using graph mincuts. In: Proceedings of the 18th International Conference on Machine Learning, pp. 19–26. Morgan Kaufmann, San Francisco (2001)
Zhu, X., Ghahramani, Z., Lafferty, J.: Semi-supervised learning using Gaussian fields and harmonic functions. In: The 20th International Conference on Machine Learning, pp. 912–919 (2003)
Zhou, D.Y., Bousquet, O., Lal, T.N., Weston, J., Schlkopf, B.: Learning with local and global consistency. In: Advances in Neural Information Processing Systems, vol. 16, pp. 321–328. MIT Press, Cambridge (2004)
Blum, A., Mitchell, T.: Combining labeled and unlabeled data with co-training. In: Proceedings of the 11th Annual Conference on Computational Learning Theory, Madison, WI, pp. 92–100 (1998)
Pierce, D., Cardie, C.: Limitations of co-training for natural language learning from large datasets. In: Proceedings of the 2001 Conference on Empirical Methods in Natural Language Processing, Pittsburgh, PA, pp. 1–9 (2001)
Wang, W., Zhou, Z.H.: Analyzing Co-training Style Algorithms. In: Kok, J.N., Koronacki, J., Lopez de Mantaras, R., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI), vol. 4701, pp. 454–465. Springer, Heidelberg (2007)
Goldman, S., Zhou, Y.: Enhancing supervised learning with unlabeled data. In: Proceedings of the 17th International Conference on Machine Learning, pp. 327–334. Morgan Kaufmann Publishers, San Francisco (2000)
Zhou, Z.H., Li, M.: Tri-training: Exploiting unlabeled data using three classifiers. IEEE Trans. on Knowledge and Data Engineering 17(11), 1529–1541 (2005)
Dietterich, T.G.: Ensemble learning. In: The Handbook of Brain Theory and Neural Networks, 2nd edn. (2002)
Dietterich, T.G.: Ensemble Methods in Machine Learning. In: Multiple Classier Systems, Cagliari, Italy (2000)
Valentini, G., Masulli, F.: Ensembles of Learning Machines. In: Marinaro, M., Tagliaferri, R. (eds.) WIRN 2002. LNCS, vol. 2486, pp. 3–19. Springer, Heidelberg (2002)
Kittler, J., Hatef, M., Duin, R.P., Matas, J.: On Combining Classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 20(3), 226–239 (1998)
Hansen, L.K., Salamon, P.: Neural network ensembles. IEEE Transaction on Pattern Analysis and Machine Intelligence 12(10), 993–1001 (1990)
Zhou, Z.H., Wu, J.X., Wei, T.: Ensembling neural networks: many could be better than all. Artificial Intelligence 137, 239–263 (2002)
Zhou, Z.H., Jiang, Y.: NeC4.5: neural ensemble based C4.5. IEEE Transactions on Knowledge and Data Engineering 16 (2004)
Breiman, L.: Bagging predictors. Machine Learning 24, 123–140 (1996)
Zhou, Z.H., Chen, S.F.: Neural Network Ensemble. Chinese J. Computers 25, 1–8 (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2012 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, X., Bai, B., Li, Y. (2012). Tri-training Based on Neural Network Ensemble Algorithm. In: Zhang, Y., Zhou, ZH., Zhang, C., Li, Y. (eds) Intelligent Science and Intelligent Data Engineering. IScIDE 2011. Lecture Notes in Computer Science, vol 7202. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-31919-8_6
Download citation
DOI: https://doi.org/10.1007/978-3-642-31919-8_6
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-31918-1
Online ISBN: 978-3-642-31919-8
eBook Packages: Computer ScienceComputer Science (R0)