Skip to main content
Log in

Generalized Mahalanobis depth in the reproducing kernel Hilbert space

  • Published:
Statistical Papers Aims and scope Submit manuscript

Abstract

In this paper, Mahalanobis depth (MHD) in the Reproducing Kernel Hilbert Space (RKHS) is proposed. First, we extend the notion of MHD to a generalized version, i.e., the generalized MHD (GHMD), to make it suitable for the small sample with singular covariance matrix. We prove that GMHD is consistent with MHD when the sample has a full-rank covariance matrix. Second, we further extend GMHD to RKHS, i.e, the kernel mapped GMHD (kmGMHD), and discuss its main properties. Numeric results show that kmGMHD can give a better depth interpretation for the sample with special shape, such as a non-convex sample set. Our proposed kmGMHD can be potentially used as a robust tool for outliers detection and data classification. In addition, we also discuss the influence of parameters on the shape of the central regions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  • Abe S (2005) Support vector machines for pattern classification, 1st edn. Springer-Verlag London Limited, Reading, pp 32–33 subsection 2.3

    Google Scholar 

  • Canu S, Grandvalet Y, Guigue V, Rakotomamonjy A (2005) Svm and kernel methods matlab toolbox. Perception Systèmes et Information, INSA de Rouen, Rouen, France

  • Chen Z, Tyler DE (2004) On the finite sample breakdown points of redescending m-estimates of location. Stat Probab Lett 69: 233–242

    Article  MathSciNet  MATH  Google Scholar 

  • Chenouri S (2004) Multivariate robust nonparametric inference based on data depth. Univ Waterloo 90: 67–89

    Google Scholar 

  • Cui X, Lin L, Yang GR (2008) An extended projection data depth and its applications to discrimination. Commun Stat Theory Methods 37(14): 2276–2290

    Article  MathSciNet  MATH  Google Scholar 

  • Donoho DL (1982) Breakdown properties of multivariate location estimators. PhD thesis, Deptment of Statistics, Harvard University

  • Gao Y (2003) Data depth based on spatial rank. Stat Probab Lett 65: 217–225

    Article  MATH  Google Scholar 

  • Ghosh AK, Chaudhuri P (2005a) On data depth and distribution-free discriminant analysis using separating surfaces. Bernoulli 11: 1–27

    Article  MathSciNet  MATH  Google Scholar 

  • Ghosh AK, Chaudhuri P (2005b) On maximum depth and related classifiers. Board Found Scand Stat 32: 327–350

    MathSciNet  MATH  Google Scholar 

  • Hofmann T, SchöLkopf B, Smola AJ (2008) Kernel methods in machine learning. Ann Stat 36(3): 1171–1220

    Article  MATH  Google Scholar 

  • Jörnsten R (2004) Clustering and classification based on the l 1 data depth. J Multivar Anal 90: 67–89

    Article  MATH  Google Scholar 

  • Li G, Zhang J (1998) Sphering and its properties. Sankhyā Indian J Stat 60(Series A, Pt. 1): 119–133

    MATH  Google Scholar 

  • Liu R, Singh K (1993) A quality index based on data depthand multivariate rank tests. J Am Stat Assoc 88: 252–260

    Article  MathSciNet  MATH  Google Scholar 

  • Liu R, Singh K (1997) Notionso of limiting p values based on data depth and bootstrap. J Am Stat Assoc 92: 266–277

    Article  MathSciNet  MATH  Google Scholar 

  • Liu RY (1990) On a notion of data depth based on random simplices. Ann Stat 18: 405–414

    Article  MATH  Google Scholar 

  • Liu RY (1992) Data depth and multivariate rank tests. In: Dodge Y (eds) L 1-Statistics and related methods. North-Holland, Amsterdam, pp 279–294

    Google Scholar 

  • Mahalaobis PC (1936) On the generalized distance in statistics. Proc Natl Acad Sci India 12: 49–55

    Google Scholar 

  • Pennacchi P (2008) Robust estimate of excitations in mechanical systems using m-estimators. J Sound Vib 310: 923–946

    Article  Google Scholar 

  • Stahel WA (1981) Robuste schatzungen: Infintesimale optimalitat and schatzungen von kovarianzmatrizen. PhD thesis, Zurich

  • Taylor JS, Cristianini N (2004) Kernel methods for pattern analysis, 1st edn. Cambridge University Press, Reading

    Book  Google Scholar 

  • Tian X, Vardi Y, Zhang C (2002) l 1−depth, depth realtive to a model, and robust regression. In: Dodge Y (eds) In Statistical data analysis based on the L 1−norm and Related Methods. Birkhauser, Basel, pp 285–299

    Google Scholar 

  • Tukey JW (1975) Mathematics and picturing of data. In: James RD (eds) Proceedings of the international congress on mathematics, vol 2. Canadian Mathematics Congress, Vancouver, pp 523–531

    Google Scholar 

  • Vardi Y, Zhang C (2000) The multivariate l 1-median and associated data depth. In: Proceedings of National Academy of Science, vol 97, pp 1423–1426

  • Zuo Y, Serfling R (2000) General notions of statistical depth function. Ann Stat 28(2): 461–482

    Article  MathSciNet  MATH  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yonggang Hu.

Additional information

This work is supported by the NSFC Grant # 60673090 and # 60802079.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Hu, Y., Wang, Y., Wu, Y. et al. Generalized Mahalanobis depth in the reproducing kernel Hilbert space. Stat Papers 52, 511–522 (2011). https://doi.org/10.1007/s00362-009-0265-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00362-009-0265-1

Keywords

Mathematics Subject Classification (2000)

Navigation