Abstract
The CoIL Challenge 2000 data mining competition attracted a wide variety of solutions, both in terms of approaches and performance. The goal of the competition was to predict who would be interested in buying a specific insurance product and to explain why people would buy. Unlike in most other competitions, the majority of participants provided a report describing the path to their solution. In this article we use the framework of bias-variance decomposition of error to analyze what caused the wide range of prediction performance. We characterize the challenge problem to make it comparable to other problems and evaluate why certain methods work or not. We also include an evaluation of the submitted explanations by a marketing expert. We find that variance is the key component of error for this problem. Participants use various strategies in data preparation and model development that reduce variance error, such as feature selection and the use of simple, robust and low variance learners like Naive Bayes. Adding constructed features, modeling with complex, weak bias learners and extensive fine tuning by the participants often increase the variance error.
Article PDF
Similar content being viewed by others
References
Berka, P. (1999). Workshop notes on discovery challenge PKDD-99.Technical report, Laboratory of Intelligent Systems, University of Economics, Prague.
Blake, C., & Merz, C. (1998). UCI Repository of machine learning databases.
Breiman, L. (1996). Bias, variance, and arcing classifiers. Technical Report, Statistics Department, University of California.
Chapman, P., Clinton, J., Khabaza, T., Reinartz, T., & Wirth, R. (1999). The CRISP-DM process model. Technical Report, Crisp Consortium. http://www.crisp-dm.org/.
Domingos, P. (1997). The role of Occam's Razor in knowledge discovery. Data Mining and Knowledge Discovery, 3, 409–425.
Domingos, P. (2000).Aunified bias-variance decomposition and its applications. In Proceedings of the Seventeenth International Conference on Machine Learning (pp. 231–238). CA: Morgan Kaufmann.
Domingos, P., & Pazzani, M. (1997). On the optimality of the simple Bayesian classifier under zero-one loss. Machine Learning, 29, 103–130.
Elkan, C., (2001). Magical thinking in data mining: Lessons from CoIL challenge 2000. In Proceedings of the Seventh International Conference on Knowledge Discovery and Data Mining (KDD'01) (pp. 426-431).
Friedman, J. (1997). On bias, variance, 0/1-loss, and the curse-of-dimensionality. Data Mining and Knowledge Discovery, 1, 55–77.
Geman, S., Bienenstock, E., & Doursat, R. (1992). Neural networks and the bias/variance dilemma. Neural Computation, 4, 1–58.
Guyon, I., & Elissee, A. (2003). An introduction to variable and feature selection. Journal of Machine Learning Research, 3, 1157–1182.
Hall, M. A. (1998). Correlation-based feature subset selection for machine learning. Ph.D. thesis, University of Waikato.
Holte, R. (1993).Very simple classification rules perform well on most commonly used datasets.Machine Learning, 11, 63–91.
James, G. M. (2003). Variance and bias for general loss functions. Machine Learning, 51, 115–135.
Kohavi, R. & John, G., (1997). Wrappers for feature subset selection. Artificial Intelligence, 97, 273–324.
Kohavi, R., & Wolpert, D. H. (1996). Bias plus variance decomposition for zero-one loss functions. In L. Saitta (Ed.), Machine learning: Proceedings of the thirteenth international conference (pp. 275–283). Morgan Kaufmann.
Quinlan, J. R., & Cameron-Jones, R. M. (1995). Oversearching and layered search in empirical learning. In IJCAI (pp. 1019–1024).
Tsamardinos, I., & Aliferis, C. (2003). Towards principled feature selection: Relevancy, filters and wrappers. In Proceedings of the Ninth International Workshop on Artificial Intelligence and Statistics.
van der Putten, P., & van Someren, M. (2000). CoiL challenge 2000: The insurance company case. Technical Report 2000-09, Leiden Institute of Advanced Computer Science, Universiteit van Leiden. http://www.liacs.nl/~putten/library/cc2000.
Witten, I., & Frank, E. (2000). Data mining: Practical machine learning tools and techniques with java implementations. San Francisco: Morgan Kaufmann Publishers.
Wolpert, D. H., & Macready, W.G. (1995).Nofree lunch theorems for search.Technical Report SFI-TR-95-02-010, The Santa Fe Institute.
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
van der Putten, P., van Someren, M. A Bias-Variance Analysis of a Real World Learning Problem: The CoIL Challenge 2000. Machine Learning 57, 177–195 (2004). https://doi.org/10.1023/B:MACH.0000035476.95130.99
Issue Date:
DOI: https://doi.org/10.1023/B:MACH.0000035476.95130.99