Skip to main content

Algorithms for Attribute Selection and Knowledge Discovery

  • Conference paper
  • First Online:
Knowledge Management in Organizations (KMO 2017)

Abstract

The features relevant selection is a task performed prior to the data mining and can be seen as one of the most important problems to solve in the data preprocessing stage an in the machine learning. With the feature selection is mainly intended to improve predictive or descriptive performance of models and implement faster and less expensive algorithms. In this paper an analysis about feature selection methods is made emphasizing on decision trees, entropy measure for ranking features, and estimation of distribution algorithms. Finally, we show the result analysis of execute the three algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Larose, D.: Data Mining: Methods and Models, pp. 1–3. Wiley-Interscience, USA (2006)

    Google Scholar 

  2. Pyle, D.: Data Preparation for Data Mining, pp. 15–19. Morgan Kaufmann Publisher, USA (1999)

    Google Scholar 

  3. Bradley, P., Mangasarian, O.: Feature selection via concave minimization and support vector machine. J. Mach. Learn. ICML, 82–90 (1998). USA

    Google Scholar 

  4. Lei, Y., Huan, L.: Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res. 5, 1205–1224 (2004). USA

    Google Scholar 

  5. Guyon, I., Elissee, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003). USA

    Google Scholar 

  6. Dash, M., Liu, H.: Feature selection for classification. J. Intell. Data Anal. 1(3), 131–156 (1996). USA

    Google Scholar 

  7. Liu, H., Lei, Y.: Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 17(4), 491–502 (2005). USA

    Article  Google Scholar 

  8. Kohavi, R., John, G.: Wrappers for feature subset selection. Artif. Intell. 97(12), 273–324 (1997). USA

    Article  Google Scholar 

  9. Jennifer, G.: Feature selection for unsupervised learning. J. Mach. Learn. Res. 5, 845–889 (2004). USA

    Google Scholar 

  10. Das, S.: Filters, wrappers and a boosting-based hybrid for feature selection. In: Proceedings of the 18th International Conference on Machine Learning, USA, pp. 74–81 (2001)

    Google Scholar 

  11. Cardie, (2001): Using decision trees to improve case-based learning. In: Utgo, P. (ed.) Proceedings of the 10th International Conference on Machine Learning, USA, pp. 25–32 (1993)

    Google Scholar 

  12. Mucciardi, A., Gose, E.: A comparison of seven techniques for choosing subsets of pattern recognition. IEEE Trans. Comput. 20, 1023–1031 (1971). USA

    Article  Google Scholar 

  13. Ruiz, R., Riquelme, J., Aguilar-Ruiz, J.: Projection-based measure for efficient feature selection. J. Intell. Fuzzy Syst. 12, 175–183 (2003). USA

    Google Scholar 

  14. Pérez, I., Sánchez, R.: Adaptación del método de reducción no lineal LLE para la selección de atributos en WEKA. In: III Conferencia Internacional en Ciencias Computacionales e Informáticas, Cuba, pp. 1–7 (2016)

    Google Scholar 

  15. Winston, P.: Inteligencia Artificial, pp. 455–460. Addison Wesley, USA (1994)

    Google Scholar 

  16. Chourasia, S.: Survey paper on improved methods of ID3 decision tree classification. Int. J. Sci. Res. Pub. 3, 1–4 (2013). USA

    Google Scholar 

  17. Rodríguez, J.: Fundamentos de minería de datos. Fondo de publicaciones de la Universidad Distrital Francisco José de Caldas, Colombia, pp. 63–64 (2010)

    Google Scholar 

  18. Changala, R., Gummadi, A., Yedukondalu, G., Raju, U.N.P.G.: Classification by decision tree induction algorithm to learn decision trees from the class-labeled training tuples. Int. J. Adv. Res. Comput. Sci. Softw. Eng. 2(4), 427–434 (2012). USA

    Google Scholar 

  19. Michell, T.: Machine Learning, pp. 50–56. McGraw Hill, USA (1997)

    Google Scholar 

  20. Han, J., Kamber, M.: Data Mining: Concepts and Techniques, vol. 3, pp. 331–336. McGraw Hill, USA (2012)

    Google Scholar 

  21. Russell, S., Norvig, P.: Artificial Intelligence: A Modern Approach, pp. 531–540. Prentice Hall, USA (2012)

    Google Scholar 

  22. Kantardzic, M.: Data Mining: Concepts, Models, Methods and Algorithms, pp. 46–48. IEEE Press Wiley-Interscience, USA (2003)

    Google Scholar 

  23. Liu, H., Motoda, H.: Feature Selection for Knowledge Discovery and Data Mining, vol. 2, pp. 30–35. Kluwer Academic Publisher, USA (2000)

    Google Scholar 

  24. Liu, H., Motoda, H.: Feature Extraction, Construction and Selection. A Data Mining Perspective, pp. 20–28. Kluwer Academic Publisher, USA (2000)

    Google Scholar 

  25. Larrañaga, P., Lozano, J.: Estimation of Distribution Algorithms. A New Tool for Evolutionary Computation, pp. 1–2. Kluwer Academic Publishers, USA (2002)

    Google Scholar 

  26. Pelikan, M., Sastry, K.: Initial-population bias in the univariate estimation of distribution algorithm. In: Proceedings of the 11th Annual Conference on Genetic and Evolutionary Computation, GECCO 2009, USA, vol. 11, pp. 429–436 (2002)

    Google Scholar 

  27. Pérez, R., Hernández, A.: Un algoritmo de estimación de distribuciones para el problema de secuencia-miento en configuración jobshop, vol. 1, pp. 1–4. Communication Del CIMAT, Mexico (2015)

    Google Scholar 

  28. Mühlenbein, H., Paaß, G.: From recombination of genes to the estimation of distributions I. Binary parameters. In: Voigt, H.-M., Ebeling, W., Rechenberg, I., Schwefel, H.-P. (eds.) PPSN 1996. LNCS, vol. 1141, pp. 178–187. Springer, Heidelberg (1996). doi:10.1007/3-540-61723-X_982

    Chapter  Google Scholar 

  29. Rodríguez, N.: Feature relevance estimation by evolving probabilistic dependency networks and weighted kernel machine. A thesis submitted to the District University Francisco José de Caldas in fulfilment of the requirements for the degree of Master of Science in Information and Communications, Colombia, pp. 3–4 (2013)

    Google Scholar 

  30. Bengoetxea, E., Larrañaga, P., Bloch, I., Perchant, A.: Estimation of distribution algorithms: a new evolutionary computation approach for graph matching problems. In: Figueiredo, M., Zerubia, J., Jain, A.K. (eds.) EMMCVPR 2001. LNCS, vol. 2134, pp. 454–469. Springer, Heidelberg (2001). doi:10.1007/3-540-44745-8_30

    Chapter  Google Scholar 

  31. Pelikan, M., Sastry, K., Cantú-Paz, E. (eds.): Scalable Optimization via Probabilistic Modeling: From Algorithms to Application. SCI, vol. 33. Springer, Heidelberg (2006). doi:10.1007/978-3-540-34954-9

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Jorge Enrique Rodríguez R. , Víctor Hugo Medina García or Lina María Medina Estrada .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Rodríguez R., J.E., García, V.H.M., Estrada, L.M.M. (2017). Algorithms for Attribute Selection and Knowledge Discovery. In: Uden, L., Lu, W., Ting, IH. (eds) Knowledge Management in Organizations. KMO 2017. Communications in Computer and Information Science, vol 731. Springer, Cham. https://doi.org/10.1007/978-3-319-62698-7_33

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-62698-7_33

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-62697-0

  • Online ISBN: 978-3-319-62698-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics