Abstract
Here is discussed what is constructive meta-learning and how it goes well compared with selective meta-learning that already becomes popular. Selective meta-learning takes multiple learning schemes with the following different ways: bagging, boosting, cascading and stacking methods. On the other hand, constructive meta-learning constructs the learning scheme proper to a given data set. We have implemented constructive meta-learning by recomposing methods into learning schemes with mining (inductive learning) method repositories that come from decomposition of popular mining algorithms. To evaluate our constructive meta-learning, we have done the comparison of the performances of our constructive meta-learning and those of two stacking methods, using UCI/ML common data sets. It has shown us that our constructive meta-learning goes better than the two stacking methods. Furthermore, it turns out to be promising that we apply constructive meta-learning to meta-learner in selective meta-learning.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Abe, H., Yamaguchi, T.: Comparing the Parallel Automatic Composition of Inductive Applications with Stacking Methods. In: Proc. of the ECML/PKDD-2003 Workshop on Parallel and Distributed Computing for Machine Learning, pp. 1–12 (2003)
Breiman, L.: Bagging Predictors. Machine Learning 24(2), 123–140 (1996)
Bernstein, A., Provost, F.: An Intelligent Assistant for Knowledge Discovery Process. In: IJICAI 2001 Workshop on Wrappers for Performance Enhancement in KDD (2001)
Blake, C.L., Merz, C.J.: UCI/ML repository of machine learning databases (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html
Booker, L.B., Holland, J.H., Goldberg, D.E.: Classifier Systems and Generic Algorithms. Artificial Intelligence 40, 235–282 (1989)
Brazdil, P., Gama, J., Henery, B.: Characterizing the Applicability of Classification Algorithms Using Meta-Level Learning. In: Bergadano, F., De Raedt, L. (eds.) ECML 1994. LNCS, vol. 784, pp. 83–102. Springer, Heidelberg (1994)
Brazdil, P., Soares, C.: A Comparison of Ranking Methods for Classification Algorithm Selection. In: Lopez de Mantaras, R., Plaza, E. (eds.) ECML 2000. LNCS (LNAI), vol. 1810, pp. 63–74. Springer, Heidelberg (2000)
Engels, R.: Planning in Knowledge Discovery in Databases; Performing Task-Oriented User-Guidance. Institute for AIFB (1996)
Frank, E., Witten, I.H.: Generating Accurate Rule Sets without Global Optimization. In: Proc. of Fifteenth International Conference on Machine Learning, pp. 144–151 (1998)
Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proc. of Thirteenth International Conference on Machine Learning, pp. 148–156 (1996)
Gama, J., Brazdil, P.: Characterization of Classification Algorithms. In: Pinto-Ferreira, C., Mamede, N.J. (eds.) EPIA 1995. LNCS, vol. 990, pp. 189–200. Springer, Heidelberg (1995)
Gama, J., Brazdil, P.: Cascade Generalization. Machine Learning 41(3), 315–343 (2000)
Giraud-Carrier, C.: Beyond predictive accuracy: what? In: Proc. of Upgrading Learning to the Meta-Level: Model Selection and Data transformation, pp. 78–85 (1998)
Hinton, G.E.: Learning distributed representations of concepts. In: Proc. of 8th Annual Conference of the Cognitive Science Society, pp. 1–12 (1986)
Kohavi, R., Sommerfield, D.: Data Mining using MLC++ | A Machine Learning Library in C++. In: Proc. 8th International Conference on Tools with Artificial Intelligence, pp. 234–245 (1996)
Metal Project: http://www.metal-kdd.org/
Michalski, R., Mozetic, I., Hong, J., Lavrac, N.: The AQ15 Inductive Learning System: An Over View and Experiments., Reports of Machine Learning and Inference Laboratory, No.MLI-86-6, George Mason University (1986)
Mitchell, T.M.: Generalization as Search. Artificial Intelligence 18(2), 203–226 (1982)
PMML (Predictive Model Markup Language): http://www.dmg.org/
Quinlan, J.R.: Induction of Decision Tree. Machine Learning 1, 81–106 (1986)
Quinlan, J.R.: Programs for Machine Learning. Morgan Kaufmann, San Francisco (1992)
Quinlan, J.R.: Bagging, Boosting and C4.5. In: Proc. of American Association for Artificial Intelligence, pp. 725–730 (1996)
Witten, I., Frank, E.: Data Mining: Practical machine learning tools and techniques with Java implementations. Morgan Kaufmann Publishers, San Francisco (2000), http://www.cs.waikato.ac.na/~ml
Wolpert, D.: Stacked Generalization. Neural Network 5(2), 241–260 (1992)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Abe, H., Yamaguchi, T. (2004). Constructive Meta-learning with Machine Learning Method Repositories. In: Orchard, B., Yang, C., Ali, M. (eds) Innovations in Applied Artificial Intelligence. IEA/AIE 2004. Lecture Notes in Computer Science(), vol 3029. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-24677-0_52
Download citation
DOI: https://doi.org/10.1007/978-3-540-24677-0_52
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22007-7
Online ISBN: 978-3-540-24677-0
eBook Packages: Springer Book Archive