Advertisement

Meta-data: Characterization of Input Features for Meta-learning

  • Ciro Castiello
  • Giovanna Castellano
  • Anna Maria Fanelli
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3558)

Abstract

Common inductive learning strategies offer the tools for knowledge acquisition, but possess some inherent limitations due to the use of fixed bias during the learning process. To overcome limitations of such base-learning approaches, a novel research trend is oriented to explore the potentialities of meta-learning, which is oriented to the development of mechanisms based on a dynamical search of bias. This could lead to an improvement of the base-learner performance on specific learning tasks, by profiting of the accumulated past experience. As a significant set of I/O data is needed for efficient base-learning, appropriate meta-data characterization is of crucial importance for useful meta-learning. In order to characterize meta-data, firstly a collection of meta-features discriminating among different base-level tasks should be identified. This paper focuses on the characterization of meta-data, through an analysis of meta-features that can capture the properties of specific tasks to be solved at base level. This kind of approach represents a first step toward the development of a meta-learning system, capable of suggesting the proper bias for base-learning different specific task domains.

Keywords

Support Vector Regression Entire Dataset Numerical Attribute Average Mutual Information Input Attribute 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Giraud-Carrier, C., Vilalta, R., Brazdil, P.: Introduction to the special issue on meta-learning. Machine Learning 54, 187–193 (2004)CrossRefGoogle Scholar
  2. 2.
    Vilalta, R., Drissi, Y.: A perspective view and survey of meta-learning. Journal of Artificial Intelligence Review 18(2), 77–95 (2002)CrossRefGoogle Scholar
  3. 3.
    Merz, C.J.: Dynamical selection of learning algorithms. In: Fisher, D., Lenz, H.J. (eds.) Learning from data: Artificial Intelligence and Statistics, Springer, New York (1995)Google Scholar
  4. 4.
    Brazdil, P.B.: Data transformation and model selection by experimentation and meta-learning. In: Nédellec, C., Rouveirol, C. (eds.) ECML 1998. LNCS, vol. 1398, pp. 11–17. Springer, Heidelberg (1998)Google Scholar
  5. 5.
    Michie, D., Spiegelhalter, D., Taylor, C.: Machine learning, neural and statistical classification. Ellis Horwood, New York (1994)zbMATHGoogle Scholar
  6. 6.
    Kalousis, A., Hilario, M.: Model selection via meta-learning: a comparative study. In: Proceedings of the 12th International IEEE Conference on Tools with AI, IEEE Press, Los Alamitos (2000)Google Scholar
  7. 7.
    Soares, C., Brazdil, P.B., Kuba, P.: A meta-learning method to select the kernel width in support vector regression. Machine Learning 54, 195–209 (2004)zbMATHCrossRefGoogle Scholar
  8. 8.
    Brazdil, P., Soares, C., Costa, J.: Ranking learning algorithms: Using IBL and meta-learning on accuracy & time results. Machine Learning 50(3), 251–277 (2003)zbMATHCrossRefGoogle Scholar
  9. 9.
    Linder, C., Studer, R.: AST: Support for Algorithm Selection with a CBR Approach. In: Proceedings of the 16th International Conference on Machine Learning, Workshop on Recent Advances in Meta-Learning and Future Work (1999)Google Scholar
  10. 10.
    Bishop, C.M.: Neural Networks for Pattern Recognition. Clarendon Press, Oxford (1995)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2005

Authors and Affiliations

  • Ciro Castiello
    • 1
  • Giovanna Castellano
    • 1
  • Anna Maria Fanelli
    • 1
  1. 1.CILab – Computational Intelligence Laboratory, Computer Science DepartmentUniversity of BariBariItaly

Personalised recommendations