Semantic Concept Classification by Joint Semi-supervised Learning of Feature Subspaces and Support Vector Machines

  • Wei Jiang
  • Shih-Fu Chang
  • Tony Jebara
  • Alexander C. Loui
Part of the Lecture Notes in Computer Science book series (LNCS, volume 5305)


The scarcity of labeled training data relative to the high-dimensionality multi-modal features is one of the major obstacles for semantic concept classification of images and videos. Semi-supervised learning leverages the large amount of unlabeled data in developing effective classifiers. Feature subspace learning finds optimal feature subspaces for representing data and helping classification. In this paper, we present a novel algorithm, Locality Preserving Semi-supervised Support Vector Machines (LPSSVM), to jointly learn an optimal feature subspace as well as a large margin SVM classifier. Over both labeled and unlabeled data, an optimal feature subspace is learned that can maintain the smoothness of local neighborhoods as well as being discriminative for classification. Simultaneously, an SVM classifier is optimized in the learned feature subspace to have large margin. The resulting classifier can be readily used to handle unseen test data. Additionally, we show that the LPSSVM algorithm can be used in a Reproducing Kernel Hilbert Space for nonlinear classification. We extensively evaluate the proposed algorithm over four types of data sets: a toy problem, two UCI data sets, the Caltech 101 data set for image classification, and the challenging Kodak’s consumer video data set for semantic concept detection. Promising results are obtained which clearly confirm the effectiveness of the proposed method.


Unlabeled Data Semantic Concept Generalize Eigenvalue Problem Locality Preserve Projection Feature Subspace 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


  1. 1.
    Joachims, T.: Transductive inference for text classification using support vector machines. In: ICML, pp. 200–209 (1999)Google Scholar
  2. 2.
    Chapelle, O., et al.: Semi-supervised learning. MIT Press, Cambridge (2006)Google Scholar
  3. 3.
    Vapnik, V.: Statistical learning theory. Wiley-Interscience, New York (1998)zbMATHGoogle Scholar
  4. 4.
    Zhu, X.: Semi-supervised learning literature survey. Computer Sciences Technique Report 1530. University of Wisconsin-Madison (2005)Google Scholar
  5. 5.
    Bengio, Y., Delalleau, O., Roux, N.: Efficient non-parametric function induction in semi-supervised learning. Technique Report 1247, DIRO. Univ. of Montreal (2004)Google Scholar
  6. 6.
    Belkin, M., Niyogi, P.: Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation 15, 1373–1396 (2003)zbMATHCrossRefGoogle Scholar
  7. 7.
    Cai, D., et al.: Spectral regression: a unified subspace learning framework for content-based image retrieval. ACM Multimedia (2007)Google Scholar
  8. 8.
    Duda, R.O., et al.: Pattern classification, 2nd edn. John Wiley and Sons, Chichester (2001)zbMATHGoogle Scholar
  9. 9.
    He, X., Niyogi, P.: Locality preserving projections. Advances in NIPS (2003)Google Scholar
  10. 10.
    Belkin, M., Niyogi, P., Sindhwani, V.: Manifold regularization: A geometric framework for learning from labeled and unlabeled examples. Journal of Machine Learning Research 7, 2399–2434 (2006)MathSciNetGoogle Scholar
  11. 11.
    Blake, C., Merz, C.: Uci repository of machine learning databases (1998),
  12. 12.
    Li, F., Fergus, R., Perona, P.: Learning generative visual models from few training examples: An incremental bayesian approach tested on 101 object categories. In: CVPR Workshop on Generative-Model Based Vision (2004)Google Scholar
  13. 13.
    Loui, A., et al.: Kodak’s consumer video benchmark data set: concept definition and annotation. In: ACM Int’l Workshop on Multimedia Information Retrieval (2007)Google Scholar
  14. 14.
    Schölkopf, B., Smola, A., Müller, K.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Computation 10, 1299–1319 (1998)CrossRefGoogle Scholar
  15. 15.
    Hsu, C., Chang, C., Lin, C.: A practical guide to support vector classification,
  16. 16.
    Lazebnik, S., Schmid, C., Ponce, J.: Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In: CVPR, vol, 2, pp. 2169–2178Google Scholar
  17. 17.
  18. 18.
    Joachims, T.: Training linear svms in linear time. ACM KDD, 217–226 (2006)Google Scholar
  19. 19.
    Fergus, R., et al.: Object class recognition by unsupervised scale-invariant learning. In: CVPR, pp. 264–271 (2003)Google Scholar
  20. 20.
    Lowe, D.: Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60, 91–110 (2004)CrossRefGoogle Scholar
  21. 21.
    Chang, S., et al.: Large-scale multimodal semantic concept detection for consumer video. In: ACM Int’l Workshop on Multimedia Information Retrieval (2007)Google Scholar
  22. 22.
    NIST TRECVID (2001 – 2007),

Copyright information

© Springer-Verlag Berlin Heidelberg 2008

Authors and Affiliations

  • Wei Jiang
    • 1
  • Shih-Fu Chang
    • 1
  • Tony Jebara
    • 1
  • Alexander C. Loui
    • 2
  1. 1.Columbia UniversityNew YorkUSA
  2. 2.Eastman Kodak CompanyRochesterUSA

Personalised recommendations