Solving Regression Problems Using Competitive Ensemble Models
- First Online:
The use of ensemble models in many problem domains has increased significantly in the last fewyears. The ensemble modeling, in particularly boosting, has shown a great promise in improving predictive performance of a model. Combining the ensemble members is normally done in a co-operative fashion where each of the ensemble members performs the same task and their predictions are aggregated to obtain the improved performance. However, it is also possible to combine the ensemble members in a competitive fashion where the best prediction of a relevant ensemble member is selected for a particular input. This option has been previously somewhat overlooked. The aim of this article is to investigate and compare the competitive and co-operative approaches to combining the models in the ensemble. A comparison is made between a competitive ensemble model and that of MARS with bagging, mixture of experts, hierarchical mixture of experts and a neural network ensemble over several public domain regression problems that have a high degree of nonlinearity and noise. The empirical results showa substantial advantage of competitive learning versus the co-operative learning for all the regression problems investigated. The requirements for creating the efficient ensembles and the available guidelines are also discussed.
Unable to display preview. Download preview PDF.
- 1.Barnett, J. A. “Computational methods for a mathematical theory of evidence”, Proceedings of IJCAI, pp. 868–875, 1981.Google Scholar
- 2.Bates, J. M. and C. W. J. Granger. “The combination of forecasts”. Operations Research Quaterly, 20:451–468, 1969.Google Scholar
- 5.Breiman, L. “Bagging predictors”. Machine Learning, 26(2):123–140, 1996.Google Scholar
- 9.Drucker, H. “Improving regressors using boosting techniques”. Proceedings of the 14th International Conference on Machine Learning, pp. 107–115, 1997.Google Scholar
- 10.Frayman, Y., Rolfe B. F., Hodgson, P. D. and Webb G. I. “Predicting the rolling force in hot steel rolling mill using an ensemble model”. Proceedings of the IASTED International Conference on Artificial Intelligence and Applications (AIA 2002), 2002. (in press).Google Scholar
- 14.Friedman, J. “Greedy function approximation: a gradient boosting machine”. Annals of Statistics, 29(4). 2001.Google Scholar
- 18.Ridgeway, G. “The state of boosting”. Computing Science and Statistics, 31:172–7181, 1999.Google Scholar
- 20.Schapire, R. E. “The strength of weak learnability”. Machine Learning, 5:197–227, 1990.Google Scholar
- 21.Sharkey, A.J.C. (Ed.) Combining artificial neural nets: ensemble and modular multi-net systems, Springer-Verlag, 1999.Google Scholar
- 22.Ting, K. M. “The characterisation of predictive accuracy and decision combination”. Proceedings of the 13th International Conference on Machine Learning, pp. 498–506, 1996.Google Scholar