Skip to main content
Log in

Multi-layer linear embedding with feature subset selection

  • Short Paper
  • Published:
Knowledge and Information Systems Aims and scope Submit manuscript

Abstract

Many fundamental problems in machine learning require some form of dimensionality reduction. To this end, two different strategies were used: manifold learning and feature selection. Manifold learning (or data embedding) attempts to compute a subspace from original data by feature recombination/transformation. Feature selection aims to select the most relevant features in the original space. In this paper, we propose a novel cooperative manifold learning-feature selection that goes beyond the simple concatenation of these two modules. Our basic idea is to learn an embedding (or the subspace) by computing a cascade of embeddings in which each embedding undergoes feature selection and elimination. We use filter approaches in order to efficiently select irrelevant features at any stage of the process. For a case study, our proposed framework was used with two typical linear embedding algorithms: local discriminant embedding (a supervised technique) and locality preserving projections (unsupervised technique) on four challenging face databases and it has been conveniently compared with other cooperative schemes. Moreover, a comparison with several state-of-the-art manifold learning methods is provided. As it is exhibited by our experimental study, the proposed framework can achieve superior learning performance with respect to classic cooperative schemes and to many competing manifold learning methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Notes

  1. http://see.xidian.edu.cn/vipsl/database_Face.html.

  2. http://vision.ucsd.edu/~leekc/ExtYaleDatabase/ExtYaleB.html.

  3. http://www.ri.cmu.edu/projects/project_418.html.

  4. http://nova.postech.ac.kr/special/imdb/imdb.html.

  5. http://www.vision.caltech.edu/Image_Datasets/Caltech101/.

References

  1. Aghazadeh RSA, LeJeune D, Dasarathy G, Shrivastava A, Baraniuk R (2018) Ultra large-scale feature selection using count-sketches. In: ICML

  2. Baudat G, Anouar F (2000) Generalized discriminant analysis using a kernel approach. Neural Comput 12(10):2385–2404

    Article  Google Scholar 

  3. Belhumeur PN, Hespanha J P, Kriegman DJ (1997) Eigenfaces vs fisherfaces: recognition using class specific linear projection. IEEE Trans Pattern Anal Mach Intell 19(7):711–720

    Article  Google Scholar 

  4. Belkin M, Niyogi P (2003) Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput 15(6):1373–1396

    Article  Google Scholar 

  5. Cai D, He X, Han J (2011) Speed up kernel discriminant analysis. Int J Very Large Data Bases 20(1):21–33

    Article  Google Scholar 

  6. Chandrashekar G, Sahin F (2014) A survey on feature selection methods. Comput Electr Eng 40(1):16–28

    Article  Google Scholar 

  7. Chen H, Chang H, Liu T (2005) Local discriminant embedding and its variants. In: IEEE international conference on computer vision and pattern recognition

  8. Dong X, Zhu L, Song X, Li J, Cheng Z (2018) Adaptive collaborative similarity learning for unsupervised multi-view feature selection. In: IJCIA

  9. Dornaika F, Assoum A (2010) Linear dimensionality reduction through eigenvector selection for object recognition. In: International symposium on visual computing, volume LNCS 6453

  10. Dornaika F, Assoum A (2013) Enhanced and parameterless locality preserving projections for face recognition. Neurocomputing 99:448–457

    Article  Google Scholar 

  11. Dornaika F, Bosaghzadeh A (2013) Exponential local discriminant embedding and its application to face recognition. IEEE Trans Cybern 43(3):921–934

    Article  Google Scholar 

  12. Du W, Cao Z, Song T, Li Y, Liang Y (2017) A feature selection method based on multiple kernel learning with expression profiles of different types. BioData Min 10(4):1–16

  13. Gilad-Bachrach R, Navot A, Tishby N (2004) Margin based feature selection—theory and algorithms. In: International conference on machine learning

  14. Guyon I, Elisseeff A (2003) An introduction to variable and feature selection. J Mach Learn Res 3:1157–1182

    MATH  Google Scholar 

  15. He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE conference on computer vision and pattern recognition (CVPR), pp 770–778

  16. Hou C, Nie F, Li X, Yi D, Wu Y (2014) Joint embedding learning and sparse regression: a framework for unsupervised feature selection. IEEE Trans Cybern 44(6):793–804

    Article  Google Scholar 

  17. Jain A, Duin R, Mao J (2000) Statistical pattern recognition: a review. IEEE Trans Pattern Anal Mach Intell 22(1):4–37

    Article  Google Scholar 

  18. Lai C, Guo S, Cheng L, Wang WA (2017) A comparative study of feature selection methods for the discriminative analysis of temporal lobe epilepsy. Front. Neurol 8(633):1–13

  19. Lai Z, Jin Z, Yang J, Wong W (2010) Sparse local discriminant projections for face feature extraction. In: International conference on pattern recognition

  20. Li J, Liu H (2017) Challenges of feature selection for big data analytics. IEEE Intell Syst 32(2):9–15

    Article  MathSciNet  Google Scholar 

  21. Nie F, Xiang S, Jia CZY, Yan S (2008) Trace ratio criterion for feature selection. In: AAAI conference on artificial intelligence

  22. Perez-Riverol Y, Kun M, Vizcaino JA, Hitz M-P, Audain E (2017) Accurate and fast feature selection workflow for high-dimensional omics data. bioRxiv

  23. Rao VM, Sastry VN (2012) Unsupervised feature ranking based on representation entropy. In: International conference on recent advances in information technology

  24. Stanczyk U, Zielosko B, Jain L (2018) Advances in feature selection for data and pattern recognition. Springer, New York

    Book  Google Scholar 

  25. Suna Z, Bebisa G, Miller R (2004) Object detection using feature subset selection. Pattern Recogn 37:2165–2176

    Article  Google Scholar 

  26. Wang F, Wang X, Zhang D, Zhang C, Li T (2009) Marginface: a novel face recognition method by average neighborhood margin maximization. Pattern Recogn 42:2863–2875

    Article  Google Scholar 

  27. Wang S, Chen H, Peng X, Zhou C (2011) Exponential locality preserving projections for small sample size problem. Neurocomputing 74(17):3654–3662

    Article  Google Scholar 

  28. Wang W, Wang R, Huang Z, Shan S, Chen X (2018) Discriminant analysis on Riemannian manifold of Gaussian distributions for face recognition with image sets. IEEE Trans Image Process 27(1):151–163

    MathSciNet  MATH  Google Scholar 

  29. Wen J, Xu Y, Li Z, Ma Z i, Xu Y (2018) Inter-class sparsity based discriminative least square regression. Neural Netw 102:36–47

    Article  Google Scholar 

  30. Yan S, Xu D, Zhang B, Zhang H, Yang Q, Lin S (2007) Graph embedding and extension: a general framework for dimensionality reduction. IEEE Trans Pattern Anal Mach Intell 29(1):40–51

    Article  Google Scholar 

  31. Yang J, Frangi A F, Yang J-Y, Zhang D, Jin Z (2005) KPCA plus LDA: a complete kernel fisher discriminant framework for feature extraction and recognition. IEEE Trans Pattern Anal Mach Intell 27(2):230–244

    Article  Google Scholar 

  32. Zhang L, Qiao L, Chen S (2010) Graph-optimized locality preserving projections. Pattern Recogn 43:1993–2002

    Article  Google Scholar 

  33. Zhang L, Zhang Q, Zhang L, Tao D, Huang X, Du B (2015) Ensemble manifold regularized sparse low-rank approximation for multiview feature embedding. Pattern Recogn 48:3102–3112

    Article  Google Scholar 

  34. Zhang P, You X, Ou W, Chen C, Cheung Y (2016) Sparse discriminative multi-manifold embedding for one-sample face identification. Pattern Recogn 52:249–259

    Article  Google Scholar 

  35. Zhang X, Gao Y (2009) Face recognition across pose: a review. Pattern Recogn 42:2876–2896

    Article  Google Scholar 

  36. Zhu X, Li X, Zhang S, Ju C, Wu X (2017) Robust joint graph sparse coding for unsupervised spectral feature selection. IEEE Trans Neural Netw Learn Syst 28(6):1263–1275

    Article  MathSciNet  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to F. Dornaika.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dornaika, F. Multi-layer linear embedding with feature subset selection. Knowl Inf Syst 63, 1029–1043 (2021). https://doi.org/10.1007/s10115-020-01535-3

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10115-020-01535-3

Keywords

Navigation