Skip to main content

Ensemble of Unsupervised and Supervised Models with Different Label Spaces

  • Conference paper
Advanced Data Mining and Applications (ADMA 2013)

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 8347))

Included in the following conference series:

  • 3192 Accesses

Abstract

Ensemble approaches of multiple supervised and unsupervised models have been exhibited to be effective in predicting labels of new instances. Those approaches, however, require the label spaces of all supervised models to be identical to the target testing instances. In many real world applications, it is often difficult to collect such supervised models for the ensemble. In contrast, it is much easier to get large amounts of supervised models with different label spaces at a stroke. In this paper, we aim to build a novel ensemble approach that allows supervised models with different label spaces. Each supervised model is associated with an anomaly detection model. We view each supervised model as a partial voter and we manage to maximize the consensus between partial voting from supervised models and unsupervised models. In the experiments, we demonstrate the effectiveness of our approach in different data sets.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Opitz, D., Maclin, R.: Popular ensemble methods: An empirical study. JAIR 11, 169–198 (1999)

    MATH  Google Scholar 

  2. Breiman, L.: Bagging predictors. Machine Learning 24(2), 123–140 (1996)

    MATH  MathSciNet  Google Scholar 

  3. Freund, Y.: Boosting a weak learning algorithm by majority. Inf. Comput. 121(2), 256–285 (1995)

    Article  MATH  MathSciNet  Google Scholar 

  4. Hoeting, J.A., Madigan, D., Raftery, A.E., Volinsky, C.T.: Bayesian model averaging: a tutorial. Statistical Science 14(4), 382–401 (1999)

    Article  MATH  MathSciNet  Google Scholar 

  5. Gao, J., Liang, F., Fan, W., Sun, Y., Han, J.: Graph-based consensus maximization among multiple supervised and unsupervised models. In: Proceedings of NIPS, pp. 585–593 (2009)

    Google Scholar 

  6. Gao, J., Fan, W., Jiang, J., Han, J.: Knowledge transfer via multiple model local structure mapping. In: KDD, pp. 283–291 (2008)

    Google Scholar 

  7. Acharya, A., Hruschka, E.R., Ghosh, J., Acharyya, S.: C3E: A framework for combining ensembles of classifiers and clusterers. In: Sansone, C., Kittler, J., Roli, F. (eds.) MCS 2011. LNCS, vol. 6713, pp. 269–278. Springer, Heidelberg (2011)

    Chapter  Google Scholar 

  8. Laurila, J.K., Gatica-Perez, D., Aad, I., Blom, J., Bornet, O., Do, T.M.T., Dousse, O., Eberle, J., Miettinen, M.: The mobile data challenge: Big data for mobile computing research. In: MCSMobile Data Challenge by Nokia Workshop, in Conjunction with International Conference on Pervasive Computing (2012)

    Google Scholar 

  9. Chandola, V., Banerjee, A., Kumar, V.: Anomaly detection: A survey. ACM Comput. Surv. 41(3) (2009)

    Google Scholar 

  10. Zhu, X.: Semi-supervised learning literature survey. Technical Report 1530, Computer Sciences, Univ. of Wisconsin-Madison (2005)

    Google Scholar 

  11. Joachims, T.: Transductive learning via spectral graph partitioning. In: ICML, pp. 290–297 (2003)

    Google Scholar 

  12. Goldberg, A.B., Zhu, X.: Seeing stars when there aren’t many stars: Graph-based semi-supervised learning for sentiment categorization. In: First Workshop Graph Based Methods for Natural Language Processing (2006)

    Google Scholar 

  13. Oza, N.C., Tumer, K.: Classifier ensembles: Select real-world applications. Information Fusion 9(1), 4–20 (2008)

    Article  Google Scholar 

  14. Wang, H., Shan, H., Banerjee, A.: Bayesian cluster ensembles. Statistical Analysis and Data Mining 4(1), 54–70 (2011)

    Article  MathSciNet  Google Scholar 

  15. Acharya, A., Hruschka, E.R., Ghosh, J., Sarwar, B.: Probabilistic combination of classifier and cluster ensembles for non-transductive learning. In: SDM (2013)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2013 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Jin, Y., Zeng, W., Zhuo, H.H., Li, L. (2013). Ensemble of Unsupervised and Supervised Models with Different Label Spaces. In: Motoda, H., Wu, Z., Cao, L., Zaiane, O., Yao, M., Wang, W. (eds) Advanced Data Mining and Applications. ADMA 2013. Lecture Notes in Computer Science(), vol 8347. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-53917-6_42

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-53917-6_42

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-53916-9

  • Online ISBN: 978-3-642-53917-6

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics