Abstract
In this hybrid decision forest each individual base decision tree classifiers are integrated with an additional classifier model, the boosted decision stump. In this boosting, observation weights for subsequent iterations are updated according to the binomial log-likelihood (L2) loss function. This boosted decision stump trained on the extra samples different than the base tree classifiers (which are defined as out-of-bag samples). This extra sample along with the subsample on which the base tree classifiers are trained approximates the original training set, so in this way we are utilizing the full training set to construct a hybrid decision forest with larger feature space. We have applied this hybrid decision forest in a real world applications: prediction of short term extreme rainfall. To check its performance we have also compared the results with relevant prediction methods of the two applications. Overall results suggest that the new hybrid decision forest is capable of yielding commendable predictive performance.
Keywords
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Hastie, T., Tibshirani, R., Friedman, J.: The elements of statistical learning: data mining, inference and prediction, 2nd edn. Springer (2009)
Dietterich, T.G.: Ensemble Methods in Machine Learning. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 1–15. Springer, Heidelberg (2000)
Kuncheva, L.I.: Combining Pattern Classifiers: Methods and Algorithms, 1st edn. Wiley-Interscience (2004)
Breiman, L.: Bagging Predictors. Machine Learning 24(2), 123–140 (1996)
Breiman, L.: Random Forests. Machine Learning 45(1), 5–32 (2001)
Hothorn, T., Lausen, B.: Double-bagging: combining classifiers by bootstrap aggregation. Pattern Recognition 36(6), 1303–1309 (2003)
Friedman, J., Hall, P.: On bagging and nonlinear estimation. Journal of Statistical Planning and Inference 137(3), 669–683 (2007)
Friedman, J., Hastie, T., Tibshirani, R.: Additive logistic regression: a statistical view of boosting. Annals of Statistics 28(2), 337–407 (2000)
Dettling, M., Buhlmann, P.: Boosting for tumor classification with gene expression data (June 2003)
Suykens, J.A.K., Vandewalle, J.: Least Squares Support Vector Machine Classifiers. Neural Processing Letters, 293–300 (1999)
Mason, I.: A model for assessment of weather forecasts. Australian Metereological Magazine 30, 291–303 (1982)
Stephenson, D.: Use of the “Odds Ratio” for Diagnosing Forecast Skill. Weather Forecasting 15(2), 221–232 (2000)
Stephenson, D.B., Casati, B., Ferro, C.A.T., Wilson, C.A.: The extreme dependency score: a non-vanishing measure for forecasts of rare events. Meteorological Applications 15(1), 41–50 (2008)
Hogan, R.J., O’Connor, E.J., Illingworth, A.J.: Verification of cloud-fraction forecasts. Quarterly Journal of the Royal Meteorological Society 135(643), 1494–1511 (2009)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag GmbH Berlin Heidelberg
About this paper
Cite this paper
Faisal, Z.M., Monira, S.S., Hirose, H. (2013). DF-ReaL2Boost: A Hybrid Decision Forest with Real L2Boosted Decision Stumps. In: Gaol, F. (eds) Recent Progress in Data Engineering and Internet Technology. Lecture Notes in Electrical Engineering, vol 156. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-28807-4_8
Download citation
DOI: https://doi.org/10.1007/978-3-642-28807-4_8
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-28806-7
Online ISBN: 978-3-642-28807-4
eBook Packages: EngineeringEngineering (R0)