Abstract
Finding optimal parameters for a model is usually a crucial task in engineering approaches to classification and modeling tasks. An automated approach is particularly desirable when a hybrid approach combining several distinct methods is to be used. In this work we present an algorithm for finding optimal parameters that works with no specific information about the underlying model and only requires the discretization of the parameter range to be considered. We will illustrate the procedure’s performance for multilayer perceptrons and support vector machines, obtaining competitive results with state-of-the-art procedures whose parameters have been tuned by experts. Our procedure is much more efficient than straight parameter search (and probably than other procedures that have appeared in the literature), but it may nevertheless require extensive computations to arrive at the best parameter values, a potential drawback that can be overcome in practice because of its highly parallelizable nature.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Cristianini, N.: An introduction to support vector machines and other kernel-based learning methods. Cambridge University Press (2000)
Duda, R., Hart P., Stork D.: Pattern classification. Wiley (2000)
Franc V., Hlaváč V.: An iterative algorithm learning the maximal margin classifier. Pattern Recognition 36, pp. 1985–1996 (2003)
Harpham C., Dawson C.W., Brown M.R.: A review of genetic algorithms applied to training radial basis function networks. Neural Comput. & Applic. 13, pp. 193–201 (2004)
Howley T., Madden M.: The genetic kernel support vector machine: description and evaluation. Artificial Intelligence Review 24, pp. 379–395 (2005)
Hsu, C.W., Chang, C.C., Lin, C.J.: A practical guide to support vector classification, www.csie.ntu.edu.tw/~cjlin/libsvmtools
Keerthi, S.S.: Efficient tuning of SVM hyperparameters using radius/margin bound and iterative algorithms. IEEE Trans. Neural Networks 13, pp. 1225–1229 (2002)
Kuo L.E, Melsheimer S.S.: Using genetic algorithms to estimate the optimum width parameter in radial basis function networks. Proceedings of the American Control Conference, Baltimore (1994)
Montgomery D.C.: Design and analysis of experiments. Wiley (1976)
Prechelt L.: Proben1: A set of neural network benchmark problems and benchmarking rules, http://digbib.ubka.uni-karlsruhe.de/eva/ira/1994/21
Schölkopf B., Smola A.J.: Learning with kernels. MIT Press (2001)
Staelin C.: Parameter selection for support vector machines. Technical Report HPL-2002-354. HP Laboratories, Israel (2002)
Statlog project datasets and results, http://www.liacc.up.pt/ML/old/statlog/datasets.html
Vijayakumar S., Wu S.: Sequential support vector classifiers and regression. Neural Computation 15, pp. 2643–2681 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Jiménez, Á.B., Lázaro, J.L., Dorronsoro, J.R. (2007). Finding Optimal Model Parameters by Discrete Grid Search. In: Corchado, E., Corchado, J.M., Abraham, A. (eds) Innovations in Hybrid Intelligent Systems. Advances in Soft Computing, vol 44. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-74972-1_17
Download citation
DOI: https://doi.org/10.1007/978-3-540-74972-1_17
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-74971-4
Online ISBN: 978-3-540-74972-1
eBook Packages: EngineeringEngineering (R0)