Advertisement

Machine Learning

, Volume 61, Issue 1–3, pp 49–69 | Cite as

Single-Class Classification with Mapping Convergence

  • Hwanjo Yu
Article

Abstract

Single-Class Classification (SCC) seeks to distinguish one class of data from universal set of multiple classes. We call the target class positive and the complement set of samples negative. In SCC problems, it is assumed that a reasonable sample of the negative data is not available. SCC problems are prevalent in the real world where positive and unlabeled data are widely available but negative data are hard or expensive to acquire. We present an SCC algorithm called Mapping Convergence (MC) that computes an accurate boundary of the target class from positive and unlabeled data (without labeled negative data). The basic idea of MC is to exploit the natural “gap” between positive and negative data by incrementally labeling negative data from the unlabeled data using the margin maximization property of SVM. We also present Support Vector Mapping Convergence (SVMC) which optimizes the MC algorithm for fast training. Our analyses show that MC and SVMC without labeled negative data significantly outperform other SCC methods. They generate as accurate boundaries as standard SVM with fully labeled data when the positive data is not very under-sampled and there exist gaps between positive and negative classes in the feature space. Our results also show that SVMC trains much faster than MC with very close accuracy.

Keywords

Single-Class Classification Support Vector Machines 

References

  1. Bileschi, S. M., & Heisele, B. (2003). Advances in component-based face detection. In IEEE Int. Workshop on Analysis and Modeling of Faces and Gestures (pp. 149–156).Google Scholar
  2. Burges, C. J. C. (1998). A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery, 2, 121–167.CrossRefGoogle Scholar
  3. Chang, C.-C., & Lin, C.-J. (2001). Training nu-support vector classifiers: Theory and algorithms. Neural Computation, 13, 2119–2147.CrossRefGoogle Scholar
  4. Christianini, N., & Shawe-Taylor, J. (2000). An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University Press.Google Scholar
  5. DeComite, F., Denis, F., & Gilleron, R. (1999). Positive and unlabeled examples help learning. In Proc. Int. Conf. Algorithmic Learning Theory (ALT'99) (pp. 219–230).Google Scholar
  6. Denis, F. (1998). PAC Learning from positive statistical queries. In Proc. Int. Conf. Algorithmic Learning Theory (ALT'99) (pp. 112–126).Google Scholar
  7. Frosini, A., Gori, M., & Priami, P. (1996). A neural network-based model for paper currency recognition and verification. IEEE Transactions on Neural Networks, 7:6, 1482–1490.CrossRefGoogle Scholar
  8. Gori, M., Lastrucci, L., & Soda, G. (1995). Autoassociator-based models for speaker verification. Pattern Recognition Letters, 17, 241–250.Google Scholar
  9. Joachims, T. (2001). A statistical learning model of text classification with support vector machines. In Proc. ACM SIGIR Int. Conf. Information Retrieval (SIGIR'01) (pp. 128–136).Google Scholar
  10. Letouzey, F., Denis, F., & Gilleron, R. (2000). Learning from positive and unlabeled examples. In Proc. Int. Conf. Algorithmic Learning Theory (ALT'00) (pp. 11–30).Google Scholar
  11. Liu, B., Lee, W. S., Yu, P. S., & Li, X. (2002). Partially supervised classification of text documents. In Proc. Int. Conf. Machine Learning (ICML'02) (pp. 387–394).Google Scholar
  12. Manevitz, L. M., & Yousef, M. (2001). One-class SVMs for document classification. Journal of Machine Learning Research, 2, 139–154.Google Scholar
  13. Scholkopf, B., Platt, J., Shawe-Taylor, J., Smola, A. J., & Williamson, R. C. (2001). Estimating the support of a high-dimensional distribution. Neural Computation, B:7 1443–1471.Google Scholar
  14. Scholkopf, B., Smola, A. J., Williamson, R. C., & Bartlett, P. L. (2000). New support vector algorithms. Neural Computation, 12, 1083–1121.Google Scholar
  15. Tax, D. M. J., & Duin, R. P. W. (2001). Uniform object generation for optimizing one-class classifiers. Journal of Machine Learning Research, 2, 155–173.Google Scholar
  16. Tong, S. and Koller, D. (2000). Support vector machine active learning with applications to text classification. In Proc. Int. Conf. Machine Learning (ICML'00) (pp. 999–1006).Google Scholar
  17. Valiant, L. G. (1984). A theory of the learnable. Communications of the ACM, 27, 1134–1142.CrossRefzbMATHGoogle Scholar
  18. Vapnik, V. N. (1998). Statistical Learning Theory. John Wiley and Sons.Google Scholar
  19. Yu, H. (2003). SVMC: Single-class classification with support vector machines. In Proc. Int. Joint Conf. on Articial Intelligence (IJCAI'03) (pp. 567–572).Google Scholar
  20. Yu, H., Han, J., & Chang, K. C. (2002). PEBL: Positive-example based learning for web page classification using SVM. In Proc. ACM SIGKDD Int. Conf. Knowledge Discovery and Data Mining (KDD'02) (pp. 239–248).Google Scholar

Copyright information

© Springer Science + Business Media, Inc. 2005

Authors and Affiliations

  1. 1.Department of Computer ScienceUniversity of IowaIowa CityUSA

Personalised recommendations