Applied Intelligence

, Volume 49, Issue 12, pp 4033–4057 | Cite as

A new hybrid feature selection based on multi-filter weights and multi-feature weights

  • Youwei WangEmail author
  • Lizhou Feng


A traditional feature selection of filters evaluates the importance of a feature by using a particular metric, deducing unstable performances when the dataset changes. In this paper, a new hybrid feature selection (called MFHFS) based on multi-filter weights and multi-feature weights is proposed. Concretely speaking, MFHFS includes the following three stages: Firstly, all samples are normalized and discretized, and the noises and the outliers are removed based on 10-folder cross validation. Secondly, the vector of multi-filter weights and the matrix of multi-feature weights are calculated and used to combine different feature subsets obtained by the optimal filters. Finally, a Q-range based feature relevance calculation method is proposed to measure the relationship of different features and the greedy searching policy is used to filter the redundant features of the temp feature subset to obtain the final feature subset. Experiments are carried out using two typical classifiers of support vector machine and random forest on six datasets (APS, Madelon, CNAE9, Gisette, DrivFace and Amazon). When the measurements of F1macro and F1micro are used, the experimental results show that the proposed method has great improvement on classification accuracy compared to the traditional filters, and it achieves significant improvements on running speed while guaranteeing the classification accuracy compared to typical hybrid feature selections.


Feature selection Feature relevance Greedy searching Support vector machine Random forest 



This research is supported by the Beijing Natural Science Foundation, China (No. 4174105), the Key Projects of National Bureau of Statistics of China (No. 2017LZ05), the National Key R&D Program of China (2017YFB1400700), the Joint Funds of the National Natural Science Foundation of China (No. U1509214).

Supplementary material

10489_2019_1470_MOESM1_ESM.m (6 kb)
ESM 1 (M 5 kb)
10489_2019_1470_MOESM2_ESM.m (2 kb)
ESM 2 (M 2 kb)
10489_2019_1470_MOESM3_ESM.m (1 kb)
ESM 3 (M 1 kb)
10489_2019_1470_MOESM4_ESM.m (2 kb)
ESM 4 (M 2 kb)
10489_2019_1470_MOESM5_ESM.m (1 kb)
ESM 5 (M 527 bytes)
10489_2019_1470_MOESM6_ESM.m (1 kb)
ESM 6 (M 726 bytes)
10489_2019_1470_MOESM7_ESM.m (1 kb)
ESM 7 (M 1 kb)
10489_2019_1470_MOESM8_ESM.m (3 kb)
ESM 8 (M 2 kb)
10489_2019_1470_MOESM9_ESM.m (0 kb)
ESM 9 (M 480 bytes)
10489_2019_1470_MOESM10_ESM.m (1 kb)
ESM 10 (M 1 kb)
10489_2019_1470_MOESM11_ESM.m (0 kb)
ESM 11 (M 386 bytes)
10489_2019_1470_MOESM12_ESM.m (5 kb)
ESM 12 (M 4 kb)
10489_2019_1470_MOESM13_ESM.m (1 kb)
ESM 13 (M 1 kb)
10489_2019_1470_MOESM14_ESM.m (1 kb)
ESM 14 (M 1 kb)
10489_2019_1470_MOESM15_ESM.m (3 kb)
ESM 15 (M 2 kb)
10489_2019_1470_MOESM16_ESM.m (1 kb)
ESM 16 (M 1 kb)
10489_2019_1470_MOESM17_ESM.m (4 kb)
ESM 17 (M 3 kb)
10489_2019_1470_MOESM18_ESM.m (2 kb)
ESM 18 (M 1 kb)
10489_2019_1470_MOESM19_ESM.m (1 kb)
ESM 19 (M 1 kb)
10489_2019_1470_MOESM20_ESM.m (1 kb)
ESM 20 (M 1 kb)
10489_2019_1470_MOESM21_ESM.m (1 kb)
ESM 21 (M 1 kb)
10489_2019_1470_MOESM22_ESM.m (2 kb)
ESM 22 (M 1 kb)
10489_2019_1470_MOESM23_ESM.m (3 kb)
ESM 23 (M 2 kb)
10489_2019_1470_MOESM24_ESM.m (11 kb)
ESM 24 (M 11 kb)
10489_2019_1470_MOESM25_ESM.m (3 kb)
ESM 25 (M 3 kb)
10489_2019_1470_MOESM26_ESM.m (3 kb)
ESM 26 (M 2 kb)
10489_2019_1470_MOESM27_ESM.m (1 kb)
ESM 27 (M 1005 bytes)
10489_2019_1470_MOESM28_ESM.m (2 kb)
ESM 28 (M 2 kb)
10489_2019_1470_MOESM29_ESM.m (1 kb)
ESM 29 (M 1 kb)
10489_2019_1470_MOESM30_ESM.m (3 kb)
ESM 30 (M 2 kb)
10489_2019_1470_MOESM31_ESM.m (1 kb)
ESM 31 (M 834 bytes)
10489_2019_1470_MOESM32_ESM.m (2 kb)
ESM 32 (M 1 kb)
10489_2019_1470_MOESM33_ESM.m (2 kb)
ESM 33 (M 1 kb)
10489_2019_1470_MOESM34_ESM.m (3 kb)
ESM 34 (M 3 kb)
10489_2019_1470_MOESM35_ESM.m (2 kb)
ESM 35 (M 1 kb)
10489_2019_1470_MOESM36_ESM.m (2 kb)
ESM 36 (M 1 kb)
10489_2019_1470_MOESM37_ESM.m (2 kb)
ESM 37 (M 1 kb)
10489_2019_1470_MOESM38_ESM.m (0 kb)
ESM 38 (M 170 bytes)
10489_2019_1470_MOESM39_ESM.m (1 kb)
ESM 39 (M 732 bytes)
10489_2019_1470_MOESM40_ESM.m (2 kb)
ESM 40 (M 2 kb)
10489_2019_1470_MOESM41_ESM.m (2 kb)
ESM 41 (M 1 kb)
10489_2019_1470_MOESM42_ESM.m (2 kb)
ESM 42 (M 2 kb)
10489_2019_1470_MOESM43_ESM.m (2 kb)
ESM 43 (M 1 kb)
10489_2019_1470_MOESM44_ESM.m (1 kb)
ESM 44 (M 1 kb)
10489_2019_1470_MOESM45_ESM.m (1 kb)
ESM 45 (M 1 kb)


  1. 1.
    Hancer E, Xue B, Zhang M (2018) Differential evolution for filter feature selection based on information theory and feature ranking. Knowledge-Based SystemsGoogle Scholar
  2. 2.
    Rawles S, Flach P (2004) Redundant feature elimination for multi-class problems. International Conference on Machine Learning ACMGoogle Scholar
  3. 3.
    Bharti KK, Singh PK (2015) Hybrid dimension reduction by integrating feature selection with feature extraction method for text clustering. Expert Syst Appl 42(6):3105–3114Google Scholar
  4. 4.
    Zabalza J, Ren J, Zheng J et al (2016) Novel segmented stacked autoencoder for effective dimensionality reduction and feature extraction in hyperspectral imaging. Neurocomputing 185(C):1–10Google Scholar
  5. 5.
    Quispe O, Ocsa A, Coronado R (2017) Latent semantic indexing and convolutional neural network for multi-label and multi-class text classification. IEEE Latin American Conference on Computational Intelligence. IEEE, 1–6Google Scholar
  6. 6.
    Marquetti I, Link JV, Lemes ALG et al (2016) Partial least square with discriminant analysis and near infrared spectroscopy for evaluation of geographic and genotypic origin of arabica coffee. Comput Electron Agric 121(C):313–319Google Scholar
  7. 7.
    Okada K, Lee MD (2016) A Bayesian approach to modeling group and individual differences in multidimensional scaling. J Math Psychol 70:35–44MathSciNetzbMATHGoogle Scholar
  8. 8.
    Fan Z, Xu Y, Zuo W et al (2017) Modified principal component analysis: an integration of multiple similarity subspace models. IEEE Transactions on Neural Networks & Learning Systems 25(8):1538–1552Google Scholar
  9. 9.
    Prihatini PM, Putra IKGD, Giriantari IAD et al (2017) Fuzzy-Gibbs latent Dirichlet allocation model for feature extraction on Indonesian documents. Contemporary Engineering Sciences 10:403–421Google Scholar
  10. 10.
    Zhang Y, Zhang Z (2012) Feature subset selection with cumulate conditional mutual information minimization. Expert Syst Appl 39(5):6078–6088Google Scholar
  11. 11.
    Yang Y, Pedersen J (1997) A comparative study on feature set selection in text categorization. In: Fisher DH (ed) Proceedings of the 14th International Conference on Machine Learning. Morgan Kaufmann, San Francisco, pp 412–420Google Scholar
  12. 12.
    Shang W, Huang H, Zhu H et al (2007) A novel feature selection algorithm for text classification. Expert Syst Appl 33(1):1–5Google Scholar
  13. 13.
    Uysal AK, Gunal S A novel probabilistic feature selection for text classification. Knowl-Based Syst 36:226–235Google Scholar
  14. 14.
    Mengle SSR, Goharian N (2009) Ambiguity measure feature-selection algorithm. J Am Soc Inf Sci Technol 60:1037–1050Google Scholar
  15. 15.
    Sebastiani F (2002) Machine learning in automated text categorization. ACM Computing Surveys (CSUR) 34(1):1–47MathSciNetGoogle Scholar
  16. 16.
    Shi JT, Liu HL, Xu Y et al (2014) Chinese sentiment classifier machine learning based on optimized information gain feature selection. Adv Mater Res 988:511–516Google Scholar
  17. 17.
    Peng H, Long F, Ding C (2005) Feature selection based on mutual information criteria of max-dependency, max-relevance, and min redundancy. IEEE Trans Pattern Anal Mach Intell 27(8):1226–1238Google Scholar
  18. 18.
    Moradi P, Rostami M (2015) Integration of graph clustering with ant colony optimization for feature selection. Knowl-Based Syst 84(C):144–161Google Scholar
  19. 19.
    Yan J, Liu N, Zhang B (2009) OCFS: optimal orthogonal centroid feature selection for text categorization. International ACM SIGIR Conference on Research and Development in Information Retrieval. ACM: 122–129Google Scholar
  20. 20.
    Yang J, Qu Z, Liu Z (2014) Improved feature-selection method considering the imbalance problem in text categorization. Sci World J:1–17Google Scholar
  21. 21.
    Tutkan M, Ganiz MC, Akyokuş S (2016) Helmholtz principle based supervised and unsupervised feature selection methods for text mining. Inf Process Manag 52(5):885–910Google Scholar
  22. 22.
    Forman G (2003) An extensive empirical study of feature selection metrics for text classification. J Mach Learn Res 3:1289–1305zbMATHGoogle Scholar
  23. 23.
    Rehman A, Javed K, Babri HA (2017) Feature selection based on a normalized difference measure for text classification. Inf Process Manag 53(2):473–489Google Scholar
  24. 24.
    Zhou X, Hu Y, Guo L (2014) Text categorization based on clustering feature selection. Procedia Computer Science 31(31):398–405Google Scholar
  25. 25.
    Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: A mutual information-based feature selection. Expert Syst Appl 41(14):6371–6385Google Scholar
  26. 26.
    Vinh LT, Lee S, Park YT et al (2012) A novel feature selection based on normalized mutual information. Appl Intell 37(1):100–120Google Scholar
  27. 27.
    Lin Y, Hu Q, Liu J et al (2015) Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 168:92–103Google Scholar
  28. 28.
    Das S (2001) Wrappers and a boosting-based hybrid for feature selection. International Conference on Machine Learning 74–81Google Scholar
  29. 29.
    Es TF, Hruschka ER, Castro LN et al (2009) A cluster-based feature selection approach. Hybrid Artificial Intelligence Systems, International Conference, Salamanca, Spain, Proceedings DBLP: 169–176Google Scholar
  30. 30.
    Jaskowiak PA, Campello RJGB (2015) A cluster based hybrid feature selection approach. Intelligent Systems. IEEE, 43–48Google Scholar
  31. 31.
    Uysal AK (2016) An improved global feature selection scheme for text classification. Expert Syst Appl 43:82–92Google Scholar
  32. 32.
    Agnihotri D (2017) Variable global feature selection scheme for automatic classification of text documents. Expert Syst Appl 81(C):268–281Google Scholar
  33. 33.
    Wang Y, Liu Y, Feng L et al (2015) Novel feature selection based on harmony search for email classification. Knowl-Based Syst 73(1):311–323Google Scholar
  34. 34.
    Zorarpacı E, Özel SA (2016) A hybrid approach of differential evolution and artificial bee colony for feature selection. Expert Syst Appl 62:91–103Google Scholar
  35. 35.
    Xue B, Zhang M, Browne WN (2014) Particle swarm optimization for feature selection in classification: novel initialization and updating mechanisms. Appl Soft Comput 18:261–276Google Scholar
  36. 36.
    Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47Google Scholar
  37. 37.
    Wang Y, Feng L (2018) Hybrid feature selection using component co-occurrence based feature relevance measurement. Expert Syst Appl 102:83–99Google Scholar
  38. 38.
    Bhattacharya S, Selvakumar S (2016) Multi-measure multi-weight ranking approach for the identification of the network features for the detection of DoS and Probe attacks. Comput J 59(6):bxv078Google Scholar
  39. 39.
    Osanaiye O, Cai H, Choo KKR et al (2016) Ensemble-based multi-filter feature selection for DDoS detection in cloud computing. EURASIP J Wirel Commun Netw 2016(1):130Google Scholar
  40. 40.
    Wang Y, Feng L, Li Y (2017) Two-step based feature selection for filtering redundant information. J Intell Fuzzy Syst 33(4):2059–2073Google Scholar
  41. 41.
    Breiman L, Friedman JH, Olshen RA (1984) Classification and regression trees. Wadsworth International Group, MonteryzbMATHGoogle Scholar
  42. 42.
    Wang Y, Feng L, Zhu J (2017) Novel artificial bee colony based feature selection for filtering redundant information. Appl Intell 3:1–18Google Scholar
  43. 43.
    Duda J (1995) Supervised and unsupervised discretization of continuous Features. Machine Learning Proceedings (2):194–202Google Scholar
  44. 44.
    Paulus J, Klapuri A (2009) Music structure analysis using a probabilistic fitness measure and a greedy search algorithm. IEEE Trans Audio Speech Lang Process 17(6):1159–1170Google Scholar
  45. 45.
    Dadaneh BZ, Markid HY, Zakerolhosseini A (2016) Unsupervised probabilistic feature selection using ant colony optimization. Expert Syst Appl 53:27–42Google Scholar
  46. 46.
    Asuncion A, Newman DJ (2007) UCI machine learning repository. University of California, Department of Information and Computer Science, IrvineGoogle Scholar
  47. 47.
    Shan S (2016) Support vector machine. Machine Learning Models and Algorithms for Big Data Classification. Springer US, 24–52Google Scholar
  48. 48.
    Breiman L (2001) Random forests. Mach Learn 45(1):5–32zbMATHGoogle Scholar
  49. 49.
    Masetic Z, Subasi A (2016) Congestive heart failure detection using random forest classifier. Comput Methods Prog Biomed 130(C):54–64Google Scholar
  50. 50.
    Chang CC, Lin CJLIBSVM (2001) A library for support vector machines. ACM Trans Intell Syst Technol 2(27):1–27Google Scholar
  51. 51.
    Chen J, Huang H, Tian S, Qu Y (2009) Feature selection for text classification with Naïve Bayes. Expert Syst Appl 36(3):5432–5435Google Scholar
  52. 52.
    Chang F, Guo J, Xu W et al (2015) A feature selection to handle imbalanced data in text classification. J Digit Inf Manag 13(3):169–175Google Scholar
  53. 53.
    Yang J, Qu Z, Liu Z (2014) Improved feature-selection method considering the imbalance problem in text categorization. Sci World J 3:625342Google Scholar
  54. 54.
    Liu WS, Chen X, Gu Q (2018) A noise tolerable feature selection framework for software defect prediction. Chinese Journal of Computers 41(3):506–520Google Scholar
  55. 55.
    Wang YW, Feng LZ (2018) A new feature selection for handling redundant information in text classification. Frontiers of Information Technology & Electronic Engineering 19(2):221–234MathSciNetGoogle Scholar

Copyright information

© Springer Science+Business Media, LLC, part of Springer Nature 2019

Authors and Affiliations

  1. 1.School of InformationCentral University of Finance and EconomicsBeijingChina
  2. 2.School of Science and EngineeringTianjin University of Finance and EconomicsTianjinChina

Personalised recommendations