Abstract
Ensemble approaches of multiple supervised and unsupervised models have been exhibited to be effective in predicting labels of new instances. Those approaches, however, require the label spaces of all supervised models to be identical to the target testing instances. In many real world applications, it is often difficult to collect such supervised models for the ensemble. In contrast, it is much easier to get large amounts of supervised models with different label spaces at a stroke. In this paper, we aim to build a novel ensemble approach that allows supervised models with different label spaces. Each supervised model is associated with an anomaly detection model. We view each supervised model as a partial voter and we manage to maximize the consensus between partial voting from supervised models and unsupervised models. In the experiments, we demonstrate the effectiveness of our approach in different data sets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Opitz, D., Maclin, R.: Popular ensemble methods: An empirical study. JAIR 11, 169–198 (1999)
Breiman, L.: Bagging predictors. Machine Learning 24(2), 123–140 (1996)
Freund, Y.: Boosting a weak learning algorithm by majority. Inf. Comput. 121(2), 256–285 (1995)
Hoeting, J.A., Madigan, D., Raftery, A.E., Volinsky, C.T.: Bayesian model averaging: a tutorial. Statistical Science 14(4), 382–401 (1999)
Gao, J., Liang, F., Fan, W., Sun, Y., Han, J.: Graph-based consensus maximization among multiple supervised and unsupervised models. In: Proceedings of NIPS, pp. 585–593 (2009)
Gao, J., Fan, W., Jiang, J., Han, J.: Knowledge transfer via multiple model local structure mapping. In: KDD, pp. 283–291 (2008)
Acharya, A., Hruschka, E.R., Ghosh, J., Acharyya, S.: C3E: A framework for combining ensembles of classifiers and clusterers. In: Sansone, C., Kittler, J., Roli, F. (eds.) MCS 2011. LNCS, vol. 6713, pp. 269–278. Springer, Heidelberg (2011)
Laurila, J.K., Gatica-Perez, D., Aad, I., Blom, J., Bornet, O., Do, T.M.T., Dousse, O., Eberle, J., Miettinen, M.: The mobile data challenge: Big data for mobile computing research. In: MCSMobile Data Challenge by Nokia Workshop, in Conjunction with International Conference on Pervasive Computing (2012)
Chandola, V., Banerjee, A., Kumar, V.: Anomaly detection: A survey. ACM Comput. Surv. 41(3) (2009)
Zhu, X.: Semi-supervised learning literature survey. Technical Report 1530, Computer Sciences, Univ. of Wisconsin-Madison (2005)
Joachims, T.: Transductive learning via spectral graph partitioning. In: ICML, pp. 290–297 (2003)
Goldberg, A.B., Zhu, X.: Seeing stars when there aren’t many stars: Graph-based semi-supervised learning for sentiment categorization. In: First Workshop Graph Based Methods for Natural Language Processing (2006)
Oza, N.C., Tumer, K.: Classifier ensembles: Select real-world applications. Information Fusion 9(1), 4–20 (2008)
Wang, H., Shan, H., Banerjee, A.: Bayesian cluster ensembles. Statistical Analysis and Data Mining 4(1), 54–70 (2011)
Acharya, A., Hruschka, E.R., Ghosh, J., Sarwar, B.: Probabilistic combination of classifier and cluster ensembles for non-transductive learning. In: SDM (2013)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jin, Y., Zeng, W., Zhuo, H.H., Li, L. (2013). Ensemble of Unsupervised and Supervised Models with Different Label Spaces. In: Motoda, H., Wu, Z., Cao, L., Zaiane, O., Yao, M., Wang, W. (eds) Advanced Data Mining and Applications. ADMA 2013. Lecture Notes in Computer Science(), vol 8347. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-53917-6_42
Download citation
DOI: https://doi.org/10.1007/978-3-642-53917-6_42
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-53916-9
Online ISBN: 978-3-642-53917-6
eBook Packages: Computer ScienceComputer Science (R0)