Advertisement

Foundation of Mining Class-Imbalanced Data

  • Da Kuang
  • Charles X. Ling
  • Jun Du
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7301)

Abstract

Mining class-imbalanced data is a common yet challenging problem in data mining and machine learning. When the class is imbalanced, the error rate of the rare class is usually much higher than that of the majority class. How many samples do we need in order to bound the error of the rare class (and the majority class)? If the misclassification cost of the class is known, can the cost-weighted error be bounded as well? In this paper, we attempt to answer those questions with PAC-learning. We derive several upper bounds on the sample size that guarantee the error on a particular class (the rare and majority class) and the cost-weighted error, with the consistent and agnostic learners. Similar to the upper bounds in traditional PAC learning, our upper bounds are quite loose. In order to make them more practical, we empirically study the pattern observed in our upper bounds. From the empirical results we obtain some interesting implications for data mining in real-world applications. As far as we know, this is the first work providing theoretical bounds and the corresponding practical implications for mining class-imbalanced data with unequal cost.

Keywords

Positive Error Cost Ratio Negative Error Class Imbalance Positive Class 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Asuncion, A., Newman, D.: UCI machine learning repository (2007), http://www.ics.uci.edu/~mlearn/mlrepository.html
  2. 2.
    Carvalho, R., Freitas, A.: A hybrid decision tree/genetic algorithm method for data mining. Inf. Sci. 163(1-3), 13–35 (2004)CrossRefGoogle Scholar
  3. 3.
    Chawla, N., Bowyer, K., Hall, L., Kegelmeyer, P.: Smote: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research 16, 321–357 (2002)zbMATHGoogle Scholar
  4. 4.
    Japkowicz, N.: Class imbalances: Are we focusing on the right issue? In: ICML-KDD 2003 Workshop: Learning from Imbalanced Data Sets (2003)Google Scholar
  5. 5.
    Klement, W., Wilk, S., Michalowski, W., Matwin, S.: Classifying Severely Imbalanced Data. In: Butz, C., Lingras, P. (eds.) Canadian AI 2011. LNCS, vol. 6657, pp. 258–264. Springer, Heidelberg (2011)CrossRefGoogle Scholar
  6. 6.
    Mitchell, T.: Machine Learning. McGraw-Hill, New York (1997)zbMATHGoogle Scholar
  7. 7.
    Ting, K.M.: The problem of small disjuncts: its remedy in decision trees. In: Proceeding of the Tenth Canadian Conference on Artificial Intelligence, pp. 91–97 (1994)Google Scholar
  8. 8.
    Valiant, L.G.: A theory of the learnable. Commun. ACM 27(11), 1134–1142 (1984)zbMATHCrossRefGoogle Scholar
  9. 9.
    Weiss, G.: Mining with rarity: a unifying framework. SIGKDD Explor. Newsl. 6(1), 7–19 (2004)CrossRefGoogle Scholar
  10. 10.
    WEKA Machine Learning Project: Weka, http://www.cs.waikato.ac.nz/~ml/weka

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Da Kuang
    • 1
  • Charles X. Ling
    • 1
  • Jun Du
    • 1
  1. 1.Department of Computer ScienceThe University of Western OntarioLondonCanada

Personalised recommendations