Skip to main content
Log in

Incremental feature selection approach to multi-dimensional variation based on matrix dominance conditional entropy for ordered data set

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

Rough set theory is a mathematical tool widely employed in various fields to handle uncertainty. Feature selection, as an essential and independent research area within rough set theory, aims to identify a small subset of important features by eliminating irrelevant, redundant, or noisy ones. In human life, data characteristics constantly change over time and other factors, resulting in ordered datasets with varying features. However, existing feature extraction methods are not suitable for handling such datasets since they do not consider previous reduction results when features change and need to be recomputed, leading to significant time consumption. To address this issue, the incremental attribute reduction algorithm utilizes prior reduction results effectively reducing computation time. Motivated by this approach, this paper investigates incremental feature selection algorithms for ordered datasets with changing features. Firstly, we discuss the dominant matrix and the dominance conditional entropy while introducing update principles for the new dominant matrix and dominance diagonal matrix when features change. Subsequently, we propose two incremental feature selection algorithms for adding (IFS-A) or deleting (IFS-D) features in ordered data set. Additionally, nine UCI datasets are utilized to evaluate the performance of our proposed algorithm. The experimental results validate that the average classification accuracy of IFS-A and IFS-D under four classifiers on twelve datasets is 82.05% and 80.75%, which increases by 5.48% and 3.68% respectively compared with the original data.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Algorithm 1
Algorithm 2
Algorithm 3
Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Data Availability

No data was used for the research described in the article.

References

  1. Lin Y, Hu Q, Liu J, Li J, Wu X (2017) Streaming feature selection for multilabel learning based on fuzzy mutual information. IEEE Trans Fuzzy Syst 25(6):1491–1507

    Google Scholar 

  2. Hu Q, Zhang L, Zhou Y, Pedrycz W (2018) Large-scale multimodality attribute reduction with multi-kernel fuzzy rough sets. IEEE Trans Fuzzy Syst 26(1):226–238

    Google Scholar 

  3. Lin Y, Li Y, Wang C, Chen J (2018) Attribute reduction for multi-label learning with fuzzy rough set. Knowl-Based Syst 152:51–61

    Google Scholar 

  4. Tan A, Wu W, Qian Y, Liang J, Chen J, Li J (2019) Intuitionistic fuzzy rough set-based granular structures and attribute subset selection. IEEE Trans Fuzzy Syst 27(3):527–539

    Google Scholar 

  5. Chen Degang, Yang Yanyan, Dong Ze (2016) An incremental algorithm for attribute reduction with variable precision rough sets. Appl Soft Comput 45:129–149

    Google Scholar 

  6. Jing Y, Li T, Luo C, Horng S, Wang G, Yu Z (2016) An incremental approach for attribute reduction based on knowledge granularity. Knowl-Based Syst 104:24–38

    Google Scholar 

  7. Yang Y, Chen D, Hui W (2017) Active sample selection based incremental algorithm for attribute reduction with rough sets. IEEE Trans Fuzzy Syst 25(4):825–838

    Google Scholar 

  8. Lang G, Cai M, Fujita H, Xiao Q (2018) Related families-based attribute reduction of dynamic covering decision information systems. Knowl-Based Syst 162:161–173

    Google Scholar 

  9. Shu W, Qian W, Xie Y (2019) Incremental approaches for feature selection from dynamic data with the variation of multiple objects. Knowl-Based Syst 163:320–331

    Google Scholar 

  10. LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD (1989) Backpropagation applied to handwritten zip code recognition. Neural Comput 1(4):541–551

    Google Scholar 

  11. Hinton GE, Osindero S, Yee YW (2006) A fast learning algorithm for deep belief nets. Neural Comput 18(7):1527–1554

    MathSciNet  Google Scholar 

  12. Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1(2):270–280

    Google Scholar 

  13. Zhao L, Hu Q, Wang WW (2015) Heterogeneous feature selection with multi-modal deep neural networks and sparse group lasso. IEEE Trans Multimed 17(11):1936–1948

    Google Scholar 

  14. Semwal VB, Mondal K, Nandi GC (2017) Robust and accurate feature selection for humanoid push recovery and classification: deep learning approach. Neural Comput Appl 28(3):565–574

    Google Scholar 

  15. Chen J, Mao Z, Zheng R, Huang Y, He L (2019) Feature selection of deep learning models for EEG-based RSVP target detection. IEICE Trans Inf Syst 102–D(4):836–844

    Google Scholar 

  16. Niu T, Wang J, Lu H, Yang W, Du P (2020) Developing a deep learning framework with two-stage feature selection for multivariate financial time series forecasting. Expert Syst Appl 148:1–17

    Google Scholar 

  17. Zou Q, Ni L, Zhang T, Wang Q (2015) Deep learning based feature selection for remote sensing scene classification. IEEE Geosci Remote Sens Lett 12(11):2321–2325

    Google Scholar 

  18. Shi H, Li H, Zhang D, Cheng C, Cao X (2018) An efficient feature generation approach based on deep learning and feature selection techniques for traffic classification. Comput Netw 132:81–98

    Google Scholar 

  19. Ding W, Lin C, Pedrycz W (2020) Multiple relevant feature ensemble selection based on multilayer co-evolutionary consensus mapreduce. IEEE Trans Cybern 50(2):425–439

    Google Scholar 

  20. Nag K, Pal NR (2020) Feature extraction and selection for parsimonious classifiers with multiobjective genetic programming. IEEE Trans Evol Comput 24(3):454–466

    Google Scholar 

  21. Labani M, Moradi P, Jalili M (2020) A multi-objective genetic algorithm for text feature selection using the relative discriminative criterion. Expert Syst Appl 149:1–21

    Google Scholar 

  22. Ma J, Gao X (2020) A filter-based feature construction and feature selection approach for classification using genetic programming. Knowl-Based Syst 196:1–14

    Google Scholar 

  23. Das AK, Das S, Ghosh A (2017) Ensemble feature selection using bi-objective genetic algorithm. Knowl-Based Syst 123:116–127

    Google Scholar 

  24. Li A, Xue B, Zhang M (2020) Multi-objective feature selection using hybridization of a genetic algorithm and direct multisearch for key quality characteristic selection. Inform Sci 523:245–265

    MathSciNet  Google Scholar 

  25. Pawlak Z (1982) Rough sets. Int J Comput Inf Sci 11(5):341–356

    Google Scholar 

  26. Sun L, Zhang X, Qian Y, Xu J, Zhang S (2019) Feature selection using neighborhood entropy-based uncertainty measures for gene expression data classification. Inform Sci 502:18–41

    MathSciNet  Google Scholar 

  27. Dai J, Hu Q, Hu H, Huang D (2018) Neighbor inconsistent pair selection for attribute reduction by rough set approach. IEEE Trans Fuzzy Syst 26(2):937–950

    Google Scholar 

  28. Wang C, Wang Y, Shao M, Qian Y, Chen D (2020) Fuzzy rough attribute reduction for categorical data. IEEE Trans Fuzzy Syst 28(5):818–830

    Google Scholar 

  29. Greco S, Matarazzo B, Slowinski R (1999) Rough approximation of a preference relation by dominance relations. European J Oper Res 117(1):63–83

    Google Scholar 

  30. Greco Salvatore, Matarazzo Benedetto, Slowinski Roman (2001) Rough sets theory for multicriteria decision analysis. European J Oper Res 129(1):1–47

    MathSciNet  Google Scholar 

  31. Błaszczyński J, Greco S, Słowiński R, Szelg M (2009) Monotonic variable consistency rough set approaches. Internat J Approx Reason 50(7):979–999

    MathSciNet  Google Scholar 

  32. Kotłowski W, Dembczyński K, Greco S, Słowiński R (2008) Stochastic dominance-based rough set model for ordinal classification. Inform Sci 178(21):4019–4037

    MathSciNet  Google Scholar 

  33. Ali A, Ali MI, Rehman N (2019) Soft dominance based rough sets with applications in information systems. Internat J Approx Reason 113:171–195

    MathSciNet  Google Scholar 

  34. Zhang X, Chen D, Tsang ECC (2017) Generalized dominance rough set models for the dominance intuitionistic fuzzy information systems. Inform Sci 378:1–25

  35. Liang J, Wang F, Dang C, Qian Y (2014) A group incremental approach to feature selection applying rough set technique. IEEE Trans Knowl Data Eng 26:294–308

    Google Scholar 

  36. Zhang X, Mei C, Chen D, Yang Y, Li J (2020) Active incremental feature selection using a fuzzy rough set-based information entropy. IEEE Trans Fuzzy Syst 28(5):901–915

    Google Scholar 

  37. Yang Y, Chen D, Hui W, Wang X (2018) Incremental perspective for feature selection based on fuzzy rough sets. IEEE Trans Fuzzy Syst 26(3):1257–1273

    Google Scholar 

  38. Yang Y, Song S, Chen D, Zhang X (2020) Discernible neighborhood counting based incremental feature selection for heterogeneous data. Int J Mach Learn Cybern 11(5):1115–1127

    Google Scholar 

  39. Shu W, Qian W, Xie Y (2020) Incremental feature selection for dynamic hybrid data using neighborhood rough set. Knowl-Based Syst 194:1–15

    Google Scholar 

  40. Liu Y, Zheng L, Xiu Y, Yin H, Zhao S, Wang X, Chen H, Li C (2020) Discernibility matrix based incremental feature selection on fused decision tables. Internat J Approx Reason 118:1–26

  41. Chen D, Dong L, Mi J (2020) Incremental mechanism of attribute reduction based on discernible relations for dynamically increasing attribute. Soft Comput 24(1):321–332

    Google Scholar 

  42. Wang F, Liang J, Qian Y (2013) Attribute reduction: A dimension incremental strategy. Knowl-Based Syst 39:95–108

    Google Scholar 

  43. Zeng A, Li T, Liu D, Zhang J, Chen H (2015) A fuzzy rough set approach for incremental feature selection on hybrid information systems. Fuzzy Sets Syst 258:39–60

    MathSciNet  Google Scholar 

  44. Wang F, Liang J, Dang C (2013) Attribute reduction for dynamic data sets. Appl Soft Comput 13(1):676–689

    Google Scholar 

  45. Wei W, Wu X, Liang J, Cui J, Sun Y (2018) Discernibility matrix based incremental attribute reduction for dynamic data. Knowl-Based Syst 140:142–157

    Google Scholar 

  46. Wei W, Song P, Liang J, Wu X (2019) Accelerating incremental attribute reduction algorithm by compacting a decision table. Int J Mach Learn Cybern 10(9):2355–2373

    Google Scholar 

  47. Cai M, Lang G, Fujita H, Li Z, Yang T (2019) Incremental approaches to updating reducts under dynamic covering granularity. Knowl-Based Syst 172:130–140

    Google Scholar 

  48. Dong L, Chen D (2020) Incremental attribute reduction with rough set for dynamic datasets with simultaneously increasing samples and attributes. Int J Mach Learn Cybern 11(6):1339–1355

    Google Scholar 

  49. Jing Y, Li T, Fujita H, Wang B, Cheng N (2018) An incremental attribute reduction method for dynamic data mining. Inform Sci 465:202–218

    MathSciNet  Google Scholar 

  50. Shannon C, Weaver W (1948) The mathematical theory of communication. Bell Syst. Tech. J. 27(3/4):373–423

    MathSciNet  Google Scholar 

  51. Hu Q, Che X, Zhang L, Zhang D, Guo M, Yu D (2012) Rank entropy based decision trees for monotonic classification. IEEE Trans Knowl Data Eng 24(11):2052–2064

    Google Scholar 

  52. Inuiguchi M, Yoshioka Y, Kusunoki Y (2009) Variable-precision dominance-based rough set approach and attribute reduction. Internat J Approx Reason 50(8):1199–1214

    MathSciNet  Google Scholar 

Download references

Acknowledgements

This work is supported by the National Natural Science Foundation of China (No. 62376229) and the Natural Science Foundation of Chongqing (No. CSTB2023NSCQ-LZX0027).

Author information

Authors and Affiliations

Authors

Contributions

Weihua Xu: Conceptualization, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation. Yifei Yang: Data curation, Methodology, Software, Visualization, Writing-original draft, Writing-review & editing. Yi Ding: Investigation, Language polishing, Methodology, Typography, Validation, Writing-review & editing. Xiyang Chen: Investigation, Methodology. Xiaofang Lv: Investigation, Methodology.

Corresponding author

Correspondence to Yi Ding.

Ethics declarations

Competing interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Xu, W., Yang, Y., Ding, Y. et al. Incremental feature selection approach to multi-dimensional variation based on matrix dominance conditional entropy for ordered data set. Appl Intell 54, 4890–4910 (2024). https://doi.org/10.1007/s10489-024-05411-3

Download citation

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10489-024-05411-3

Keywords

Navigation