Skip to main content
Log in

A stacking ensemble learning for ship fuel consumption prediction under cross-training

  • Original Article
  • Published:
Journal of Mechanical Science and Technology Aims and scope Submit manuscript

Abstract

Accurate ship fuel consumption prediction is vital for the shipping industry. In this study, a stacking ensemble learning is developed to predict tanker fuel consumption precisely, built on cross-training of the first-level learner. Among comparative experiments, stacking with Bayesian regression (BR) as the meta-learner and extremely randomized trees (ET), gradient boosting decision tree (GBDT) and light gradient boosting machine (LGBM) as firstlevel learners achieves superior performance, yielding the best results. The root mean square error (RMSE) on the test dataset is 0.2679, and on the training dataset is 0.1327. Ensemble model-based feature importance analysis reveals that ship attributes (speed, draught, trim) contribute around 80 %, while meteorological features contribute about 20 %.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

Abbreviations

BR :

Bayesian regression

ET :

Extremely randomized trees

GBDT :

Gradient boosting decision tree

LGBM :

Light gradient boosting machine

RMSE :

Root mean square error

IMO :

International Maritime Organization

ML :

Machine learning

MLR :

Multiple linear regression

RR :

Ridge regression

ANN :

Artificial neural network

SVM :

Support vector machine

DT :

Decision tree

KNN :

K-nearest neighbor

RF :

Random forest

AB :

Adaptive boosting

TPE :

Tree-structured parzen

XGB :

Extreme gradient boosting

R 2 :

R square

EI :

Expected incremental

GC :

Golden creation

References

  1. C. Dere and C. Deniz, Load optimization of central cooling system pumps of a container ship for the slow steaming conditions to enhance the energy efficiency, Journal of Cleaner Production, 222 (2019) 206–217.

    Article  Google Scholar 

  2. UN, Review of Maritime Transport, United Nations Publications, New York, USA (2021).

    Google Scholar 

  3. IMO, Fourth Greenhouse Gas Study 2020, IMO, https://www.imo.org/en/OurWork/Environment/Pages/Fourth-IMO-Greenhouse-Gas-Study-2020.aspx.

  4. Q. Meng, Y. Zhang and M. Xu, Viability of transarctic shipping routes: A literature review from the navigational and commercial perspectives, Maritime Policy and Management, 44(1) (2017) 16–41.

    Article  Google Scholar 

  5. R. Adland et al., Does fuel efficiency pay? Empirical evidence from the drybulk timecharter market revisited, Transportation Research Part A: Policy and Practice, 95 (2017) 1–12.

    Google Scholar 

  6. Y. Du et al., Two-phase optimal solutions for ship speed and trim optimization over a voyage using voyage report data, Transportation Research Part B: Methodological, 122 (2019) 88–114.

    Article  Google Scholar 

  7. Y. Du et al., Data fusion and machine learning for ship fuel efficiency modeling: part II - voyage report data, AIS data and meteorological data, Communications in Transportation Research, 2 (2022) 100073.

    Article  Google Scholar 

  8. O. Sagi and L. Rokach, Ensemble learning: A survey, WIREs Data Mining and Knowledge Discovery, 8(4) (2018) e1249.

    Article  Google Scholar 

  9. D. Berrar, Cross-validation, Encyclopedia of Bioinformatics and Computational Biology, Elsevier, 1 (2018) 542–545.

    Google Scholar 

  10. F. Hutter, L. Kotthoff and J. Vanschoren (eds), Automated Machine Learning: Methods, Systems, Challenges, Springer International Publishing, Cham (2019).

    Google Scholar 

  11. B. Shahriari et al., Taking the human out of the loop: A review of bayesian optimization, Proceedings of the IEEE, 104(1) (2016) 148–175.

    Article  Google Scholar 

  12. M. Simonsen, H. J. Walnum and S. Gössling, Model for estimation of fuel consumption of cruise ships, Energies, 11(5) (2018) 1059.

    Article  Google Scholar 

  13. O. Soner, E. Akyuz and M. Celik, Statistical modelling of ship operational performance monitoring problem, J. Mar. Sci. Technol., 24(2) (2019) 543–552.

    Article  Google Scholar 

  14. Y.-R. Kim, M. Jung and J.-B. Park, Development of a fuel consumption prediction model based on machine learning using ship in-service data, Journal of Marine Science and Engineering, 9(2) (2021) 137.

    Article  Google Scholar 

  15. Y. B. A. Farag and A. I. Ölçer, The development of a ship performance model in varying operating conditions based on ANN and regression techniques, Ocean Engineering, 198 (2020) 106972.

    Article  Google Scholar 

  16. M. Jeon et al., Prediction of ship fuel consumption by using an artificial neural network, J. Mech. Sci. Technol., 32(12) (2018) 5785–5796.

    Article  Google Scholar 

  17. T. Uyanık, Ç. Karatuğ and Y. Arslanoğlu, Machine learning approach to ship fuel consumption: A case of container vessel, Transportation Research Part D: Transport and Environment, 84 (2020) 102389.

    Article  Google Scholar 

  18. X. Dong et al., A survey on ensemble learning, Front. Comput. Sci., 14(2) (2020) 241–258.

    Article  Google Scholar 

  19. I. Ullah et al., Electric vehicle energy consumption prediction using stacked generalization: an ensemble learning approach, International Journal of Green Energy, 18(9) (2021) 896–909.

    Article  Google Scholar 

  20. Z. Hu et al., A two-step strategy for fuel consumption prediction and optimization of ocean-going ships, Ocean Engineering, 249 (2022) 110904.

    Article  Google Scholar 

  21. S. Cui et al., A stacking-based ensemble learning method for earthquake casualty prediction, Applied Soft Computing, 101 (2021) 107038.

    Article  Google Scholar 

  22. M. H. D. M. Ribeiro et al., Efficient bootstrap stacking ensemble learning model applied to wind power generation forecasting, International Journal of Electrical Power and Energy Systems, 136 (2022) 107712.

    Article  Google Scholar 

  23. Q. Li and Z. Song, Prediction of compressive strength of rice husk ash concrete based on stacking ensemble learning model, Journal of Cleaner Production, 382 (2023) 135279.

    Article  Google Scholar 

  24. C. M. Bishop, Pattern Recognition and Machine Learning, Springer (2006).

  25. L. Breiman, Random forests, Machine Learning, 45(1) (2001) 5–32.

    Article  Google Scholar 

  26. P. Geurts, D. Ernst and L. Wehenkel, Extremely randomized trees, Mach Learn, 63(1) (2006) 3–42.

    Article  Google Scholar 

  27. C. M. Bishop and N. M. Nasrabadi, Pattern Recognition and Machine Learning, Springer, New York, 4 (4) (2006).

    Google Scholar 

  28. J. H. Friedman, Greedy function approximation: A gradient boosting machine, The Annals of Statistics, 29(5) (2001) 1189–1232.

    Article  MathSciNet  Google Scholar 

  29. T. Chen and C. Guestrin, XGBoost: A scalable tree boosting system, Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, USA (2016) 785–794.

  30. G. Ke et al., LightGBM: A highly efficient gradient boosting decision tree, Advances in Neural Information Processing Systems, Curran Associates, Inc. (2017).

  31. Skitit-learn.org, Scikit-learn: Machine Learning in Phthon-Scikit-learn 122 Documentation, https://scikit-learn.org/stable/index.html.

  32. XGBoost, XGBoost Documentation-XGBoost 174 Documentation, XGBoost, https://xgboost.readthedocs.io/en/stable/.

  33. LightGBM, Welcome to LightGBM’s Documantation-LightGBM 33599 Documentation, LightGBM, https://lightgbm.readthedocs.io/en/latest/index.html.

  34. C. Perlich, IBM Research Report: Learning Curves in Machine Learning, IBM (2009).

  35. A. Singh, A comprehensive guide to ensemble learning (with Python codes), Analytics Vidhya (2018) https://www.analytics-vidhya.com/blog/2018/06/comprehensive-guide-for-ensemble-models/.

  36. A. I. Naimi and L. B. Balzer, Stacked generalization: an introduction to super learning, Eur. J. Epidemiol., 33(5) (2018) 459–464.

    Article  Google Scholar 

  37. P. Liashchynskyi and P. Liashchynskyi, Grid search, random search, genetic algorithm: A big comparison for NAS, arXiv: 192.06059 (2019).

  38. J. Snoek, H. Larochelle and R. P. Adams, Practical bayesian optimization of machine learning algorithms, arXiv:1206.2944 (2012).

  39. A. H. Victoria and G. Maragatham, Automatic tuning of hyperparameters using Bayesian optimization, Evolving Systems, 12(1) (2021) 217–223.

    Article  Google Scholar 

  40. L. Breiman, Classification and Regression Trees, Routledge, New York (2017).

    Book  Google Scholar 

  41. Y. Freund and R. E. Schapire, A decision-theoretic generali zation of on-line learning and an application to boosting, Journal of Computer and System Sciences, 55(1) (1997) 119–139.

    Article  MathSciNet  Google Scholar 

  42. C. M. Bishop and M. E. Tipping, Bayesian regression and classification, J. A. K. Suykens, I. Horvath, S. Basu, C. Micchelli and JV (eds.), Advances in Learning Theory: Methods, Models and Applications. NATO Science Series, III: Computer and Systems Sciences, IOS Press, 190 (2003) 267–285.

  43. H. Hersbach et al., ERA5 Hourly Data on Single Levels from 1959 to Present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) (2018).

  44. ECMWF, ORAS5 Global Ocean Reanalysis Monthly Data from 1958 to Present, Copernicus Climate Change Service (C3S) Climate Data Store (CDS) (2021).

  45. M. H. Rio et al., Beyond GOCE for the ocean circulation estimate: synergetic use of altimetry, gravimetry, and in situ data provides new insight into geostrophic and Ekman currents, Geophys. Res. Lett., 41(24) (2014) 8918–8925.

    Article  Google Scholar 

Download references

Acknowledgments

This work is supported by the National Social Science Fund (21BGJ073) of China.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zhuo Sun.

Additional information

Mengjie Ma is a graduate student of the College of Transportation Engineering, Dalian Maritime University, Dalian, Liaoning, China. His research interests include green and intelligent transportation, machine learning, data mining and operational research.

Zhuo Sun is a Professor at the Department of Logistics in Dalian Maritime University. His research interests are in the areas of shipping network design and port planning. He developed a spatial planning tool named MicroCity (https://microcity.github.io).

Peixiu Han is a doctoral student of the College of Transportation Engineering, Dalian Maritime University, Dalian, Liaoning, China. Her research interests include green and intelligent transportation, machine learning, data mining and operational research.

Huirong Yang is a graduate student of the College of Transportation Engineering, Dalian Maritime University, Dalian, Liaoning, China. Her research interests include green and intelligent transportation, machine learning, data mining and operational research.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ma, M., Sun, Z., Han, P. et al. A stacking ensemble learning for ship fuel consumption prediction under cross-training. J Mech Sci Technol 38, 299–308 (2024). https://doi.org/10.1007/s12206-023-1224-9

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12206-023-1224-9

Keywords

Navigation