Skip to main content
Log in

Hybrid method to supervise feature selection using signal processing and complex algebra techniques

  • Published:
Multimedia Tools and Applications Aims and scope Submit manuscript

Abstract

Research in AI has proved to be revolutionarily beneficial to humankind from the past few decades. Many supporting techniques have been developed that indirectly evolved AI and directly enhanced various machine learning models; one of them being the feature engineering. It can also be considered as applied-ML. Another is the so-called feature selection which is method in which most contributing feature to final decision making, out of the entire feature space are selected for processing into an ML model. It is not an easy task to precisely calculate the dependency of the output variable onto the candidate features, particularly when the data is high in dimensions. In this regard, this study proposes a novel method named the cisoidal analysis based feature selection (CAFS) which uses both pre-established algorithms as well as a new approach of relating members of feature space to a complex sinusoid (cisoid) mathematically, then using signal processing techniques to eliminate certain elements in the entire feature space for enhanced feature selection and hence to obtain higher classification accuracy. Derived from experiments with five high dimensional datasets, CAFS displays significantly competitive performance than some of the pre-existing algorithms. CAFS is highly advantageous in reducing dimension of feature space in most of the applications.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11

Similar content being viewed by others

References

  1. Bach FR (2008) Bolasso: model consistent LASSO estimation through the bootstrap. In: Proceedings of the 25th international conference on machine learning, 5 July 2008, pp 33–40

  2. Battiti R (1994) Using mutual information for selecting features in supervised neural net learning. IEEE Trans Neural Netw 5(4):537–550

    Article  Google Scholar 

  3. Bermingham ML, Pong-Wong R, Spiliopoulou A, Hayward C, Rudan I, Campbell H, Wright AF, Wilson JF, Agakov F, Navarro P, Haley CS (2015) Application of high-dimensional feature selection: evaluation for genomic prediction in man. Sci Rep 5:10312

    Article  Google Scholar 

  4. Brown G, Pocock A, Zhao MJ, Luján M (2012) Conditional likelihood maximisation: a unifying framework for information theoretic feature selection. J Mach Learn Res 13(1):27–66

    MathSciNet  MATH  Google Scholar 

  5. Forman G (2003) An extensive empirical study of feature selection metrics for text classification. J Mach Learn Res 3:1289–1305

    MATH  Google Scholar 

  6. Gao W, Hu L, Zhang P (2018) Class-specific mutual information variation for feature selection. Pattern Recognit 79:328–339

    Article  Google Scholar 

  7. Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47. https://doi.org/10.1016/j.eswa.2015.12.004

    Article  Google Scholar 

  8. Guyon I, Elisseeff A (2003) An introduction to variable and feature selection. J Mach Learn Res 3:1157–1182

    MATH  Google Scholar 

  9. Guyon I, Weston J, Barnhill S, Vapnik V (2002) Gene selection for cancer classification using support vector machines. Mach Learn 46(1–3):389–422

    Article  MATH  Google Scholar 

  10. Guyon I, Weston J, Barnhill S, Vapnik V (2002) Gene selection for cancer classification using support vector machines. Mach Learn 46(1):389–422. https://doi.org/10.1023/a:1012487302797

    Article  MATH  Google Scholar 

  11. Han K, Wang Y, Zhang C, Li C, Xu C (2018) Autoencoder inspired unsupervised feature selection. In: 2018 IEEE international conference on acoustics, speech and signal processing (ICASSP), 15 April 2018. IEEE, pp 2941–2945

  12. Hsu HH, Hsieh CW, Lu MD (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38(7):8144–8150. https://doi.org/10.1016/j.eswa.2010.12.156

    Article  Google Scholar 

  13. Jain I, Jain VK, Jain R (2018) Correlation feature selection based improved-binary particle swarm optimization for gene selection and cancer classification. Appl Soft Comput 62:203–215. https://doi.org/10.1016/j.asoc.2017.09.038

    Article  Google Scholar 

  14. Jordan C (1983) Cours d'Analyse de l'École Polytechnique, vol II, Calcul Intégral: Intégrales définies et indéfinies, 2nd edn. Paris

  15. Ke Y, Zhang D (2015) Feature selection and analysis on correlated gas sensor data with recursive feature elimination. Sens Actuators B 212:353–363. https://doi.org/10.1016/j.snb.2015.02.025

    Article  Google Scholar 

  16. Kwak N, Choi CH (2002) Input feature selection by mutual information based on Parzen window. IEEE Trans Pattern Anal Mach Intell 24(12):1667–1671

    Article  Google Scholar 

  17. Lewis DD (1992) Feature selection and feature extract ion for text categorization. In: Speech and natural language: proceedings of a workshop held at Harriman, New York, 23–26 February 1992

  18. Li J, Cheng K, Wang S, Morstatter F, Trevino RP, Tang J, Liu H (2017) Feature selection: a data perspective. ACM Comput Surv 50(6):1–45

    Article  Google Scholar 

  19. Lin D, Tang X (2006) Conditional infomax learning: an integrated framework for feature extraction and fusion. In: European conference on computer vision, 7 May 2006. Springer, Berlin, pp 68–82

  20. Ma J, Teng G (2019) A hybrid multiple feature construction approach for classification using genetic programming. Appl Soft Comput 80:687–699. https://doi.org/10.1016/j.asoc.2019.04.039

    Article  Google Scholar 

  21. Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238

    Article  Google Scholar 

  22. Saeys Y, Inza I, Larrañaga P (2007) A review of feature selection techniques in bioinformatics. Bioinformatics 23(19):2507–2517

    Article  Google Scholar 

  23. Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27(3):379–423

    Article  MathSciNet  MATH  Google Scholar 

  24. Soni R, Kumar B, Chand S (2019) Optimal feature and classifier selection for text region classification in natural scene images using Weka tool. Multimed Tools Appl 78:31757–31791. https://doi.org/10.1007/s11042-019-07998-z

    Article  Google Scholar 

  25. Urbanowicz RJ, Meeker M, La Cava W, Olson RS, Moore JH (2018) Relief-based feature selection: introduction and review. J Biomed Inform 85:189–203

    Article  Google Scholar 

  26. Wei G, Zhao J, Feng Y, He A, Yu J (2020) A novel hybrid feature selection method based on dynamic feature importance. Appl Soft Comput 93:106337

    Article  Google Scholar 

  27. Xu J, Tang B, He H, Man H (2016) Semisupervised feature selection based on relevance and redundancy criteria. IEEE Trans Neural Netw Learn Syst 28(9):1974–1984

    Article  MathSciNet  Google Scholar 

  28. Yang Y, Pedersen JO (1997) A comparative study on feature selection in text categorization. In: ICML 1997, 8 July 1997, vol 97(412–420), p 35

  29. Zeng Z, Zhang H, Zhang R, Yin C (2015) A novel feature selection method considering feature interaction. Pattern Recognit 48(8):2656–2666

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Shubham Mahajan.

Ethics declarations

Conflict of interest

Authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Mahajan, S., Pandit, A.K. Hybrid method to supervise feature selection using signal processing and complex algebra techniques. Multimed Tools Appl 82, 8213–8234 (2023). https://doi.org/10.1007/s11042-021-11474-y

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11042-021-11474-y

Keywords

Navigation