Skip to main content

Decision-Rule Solutions for Data Mining with Missing Values

  • Conference paper

Part of the Lecture Notes in Computer Science book series (LNAI,volume 1952)

Abstract

A method is presented to induce decision rules from data with missing values where (a) the format of the rules is no different than rules for data without missing values and (b) no special features are spe- cified to prepare the the original data or to apply the induced rules. This method generates compact Disjunctive Normal Form (DNF) rules. Each class has an equal number of unweighted rules. A new example is classi- fied by applying all rules and assigning the example to the class with the most satisfied rules. Disjuncts in rules are naturally overlapping. When combined with voted solutions, the inherent redundancy is enhanced. We provide experimental evidence that this transparent approach to classi- fication can yield strong results for data mining with missing values.

Keywords

  • decision rule induction
  • boosting

This is a preview of subscription content, access via your institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • DOI: 10.1007/3-540-44399-1_1
  • Chapter length: 10 pages
  • Instant PDF download
  • Readable on all devices
  • Own it forever
  • Exclusive offer for individuals only
  • Tax calculation will be finalised during checkout
eBook
USD   109.00
Price excludes VAT (USA)
  • ISBN: 978-3-540-44399-5
  • Instant PDF download
  • Readable on all devices
  • Own it forever
  • Exclusive offer for individuals only
  • Tax calculation will be finalised during checkout
Softcover Book
USD   139.00
Price excludes VAT (USA)

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. E. Bauer and R. Kohavi. An empirical comparison of voting classification algorithms: Bagging, boosting and variants. Machine Learning, 36(1): 105–139, 1999.

    CrossRef  Google Scholar 

  2. C. Blake, E. Keogh, and C. Merz. Uci repository of machine learning databases. Technical report, University of California Irvine, 1999. http://www.ics.uci.edu/~mlearn/MLRepository.html

  3. L. Breiman. Bagging predictors. Machine Learning, 24:123–140, 1996.

    MATH  MathSciNet  Google Scholar 

  4. L. Breiman, J. Friedman, R. Olshen, and C. Stone. Classification and Regression Trees. Wadsworth, Monterrey, CA., 1984.

    Google Scholar 

  5. W. Cohen. Fast effective rule induction. In Proceedings of the Twelfth International Conference on Machine Learning, pages 115–123, 1995.

    Google Scholar 

  6. J. Friedman, T. Hastie, and R. Tibshirani. Additive logistic regression: A statistical view of boosting. Technical report, Stanford University Statistics Department, 1998. http://www.stat-stanford.edu/~tibs

  7. D. Pyle. Data Preparation for Data Mining. Morgan Kaufmann, San Francisco, 1999.

    Google Scholar 

  8. J. Quinlan. Unknown attribute values in induction. In International Workshop on Machine Learning, pages 164–168, Ithica, NY, 1989.

    Google Scholar 

  9. R. Schapire. A brief introduction to boosting. In Proceedings of International Joint Conference on Artificial Intelligence, pages 1401–1405, 1999.

    Google Scholar 

  10. S. Weiss, C. Apté, F. Damerau, and et al. Maximizing text-mining performance. tiIEEE Intelligent Systems, 14(4): 63–69, 1999.

    Google Scholar 

  11. S. Weiss and N. Indurkhya. Optimized rule induction. IEEE EXPERT, 8(6): 61–69, December 1993.

    CrossRef  Google Scholar 

  12. S. Weiss and N. Indurkhya. Predictive Data Mining: A Practical Guide. Morgan Kaufmann, 1998. DMSK Software: http://www.data-miner.com

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and Permissions

Copyright information

© 2000 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Weiss, S.M., Indurkhya, N. (2000). Decision-Rule Solutions for Data Mining with Missing Values. In: Monard, M.C., Sichman, J.S. (eds) Advances in Artificial Intelligence. IBERAMIA SBIA 2000 2000. Lecture Notes in Computer Science(), vol 1952. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44399-1_1

Download citation

  • DOI: https://doi.org/10.1007/3-540-44399-1_1

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-41276-2

  • Online ISBN: 978-3-540-44399-5

  • eBook Packages: Springer Book Archive