Skip to main content

A Scalable Algorithm for Learning a Mahalanobis Distance Metric

  • Conference paper
Computer Vision – ACCV 2009 (ACCV 2009)

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 5996))

Included in the following conference series:

Abstract

A distance metric that can accurately reflect the intrinsic characteristics of data is critical for visual recognition tasks. An effective solution to defining such a metric is to learn it from a set of training samples. In this work, we propose a fast and scalable algorithm to learn a Mahalanobis distance. By employing the principle of margin maximization to secure better generalization performances, this algorithm formulates the metric learning as a convex optimization problem with a positive semidefinite (psd) matrix variable. Based on an important theorem that a psd matrix with trace of one can always be represented as a convex combination of multiple rank-one matrices, our algorithm employs a differentiable loss function and solves the above convex optimization with gradient descent methods. This algorithm not only naturally maintains the psd requirement of the matrix variable that is essential for metric learning, but also significantly cuts down computational overhead, making it much more efficient with the increasing dimensions of feature vectors. Experimental study on benchmark data sets indicates that, compared with the existing metric learning algorithms, our algorithm can achieve higher classification accuracy with much less computational load.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Winn, J., Criminisi, A., Minka, T.: Object categorization by learned universal visual dictionary. In: Proc. IEEE International Conference on Computer Vision, Beijing, China, pp. 1800–1807 (2005)

    Google Scholar 

  2. Smeulders, A.W.M., Worring, M., Santini, S., Gupta, A., Jain, R.: Content-based image retrieval at the end of the early years. IEEE Transactions on Pattern Analysis and Machine Intelligence 22(12), 1349–1380 (2000)

    Article  Google Scholar 

  3. Cover, T., Hart, P.: Nearest neighbor pattern classification. IEEE Transactions on Information Theory 13, 21–27 (1967)

    Article  MATH  Google Scholar 

  4. Scholkpf, B., Smola, A.: Learning with Kernels, Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, Cambridge (2002)

    Google Scholar 

  5. Xing, E.P., Ng, A.Y., Jordan, M.I., Russell, S.: Distance metric learning, with application to clustering with side-information. In: Proc. Advances in Neural Information Processing Systems, pp. 505–512. MIT Press, Cambridge (2003)

    Google Scholar 

  6. Yang, L., Sukthankar, R., Hoi, S.C.: A boosting framework for visuality-preserving distance metric learning and its application to medical image retrieval. IEEE Transactions on Pattern Analysis and Machine Intelligence 10 (November 2008)

    Google Scholar 

  7. Vapnik, V.: Statistical learning theory. John Wiley and Sons Inc., New York (1998)

    MATH  Google Scholar 

  8. Shen, C., Welsh, A., Wang, L.: PSDBoost: Matrix-generation linear programming for positive semidefinite matrices learning. In: Proc. Advances in Neural Information Processing Systems, pp. 1473–1480. MIT Press, Cambridge (2008)

    Google Scholar 

  9. Weinberger, K.Q., Blitzer, J., Saul, L.K.: Distance metric learning for large margin nearest neighbor classification. In: Proc. Advances in Neural Information Processing Systems, pp. 1475–1482 (2006)

    Google Scholar 

  10. Goldberger, J., Roweis, S., Hinton, G., Salakhutdinov, R.: Neighbourhood components analysis. In: Proc. Advances in Neural Information Processing Systems (2005)

    Google Scholar 

  11. Nocedal, J., Wright, S.J.: Numerical optimization. Springer, New York (1999)

    Book  MATH  Google Scholar 

  12. Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental bayesian approach tested on 101 object categories. In: Workshop on Generative-Model Based Vision, IEEE Conference on Computer Vision and Pattern Recognition (2004)

    Google Scholar 

  13. Hsu, C.W., Lin, C.J.: A comparison of methods for multi-class support vector machines. IEEE Transactions on Neural Networks 13(2), 415–425 (2002)

    Article  Google Scholar 

  14. Mikolajczyk, K., Schmid, C.: Scale & affine invariant interest point detectors. International Journal of Computer Vision 60(1), 63–86 (2004)

    Article  Google Scholar 

  15. Lowe, D.G.: Object recognition from local scale-invariant features. In: Proc. IEEE International Conference on Computer Vision, pp. 1150–1157 (1999)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2010 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Kim, J., Shen, C., Wang, L. (2010). A Scalable Algorithm for Learning a Mahalanobis Distance Metric. In: Zha, H., Taniguchi, Ri., Maybank, S. (eds) Computer Vision – ACCV 2009. ACCV 2009. Lecture Notes in Computer Science, vol 5996. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-12297-2_29

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-12297-2_29

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-12296-5

  • Online ISBN: 978-3-642-12297-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics