Abstract
This chapter is concerned with the schemes of combining multiple machine learning algorithms with each other. We mention the four main combination schemes, depending on the role of the coordinator as the start of this chapter. Afterward, we describe the meta-learning for reinforcing the combination schemes. We present some partitions for implementing the ensemble learning. This chapter focuses on the combination of supervised learning algorithms for classifying data items. This chapter is intended to describe the ensemble learning, together with the meta-learning and the partition scheme.
In Sect. 13.1, we introduce the ensemble learning by providing its overviews and in Sect. 13.2, mention the schemes of combining machine learning algorithms. In Sect. 13.3, we devote to the meta-learning which is applicable to the coordinator of the machine learning algorithms. In Sect. 13.4, we mention the input partition, the architecture partition, and the training set partition which are necessary for the ensemble learning, and in Sect. 13.5, we make the summarization on this chapter and the further discussions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
T. Jo, Text classification using feature similarity based K nearest neighbor. AS Med. Sci. 3(4), 13–21 (2019)
H. Mohanty, P. Bhuyan, D. Chenthati, Big Data: A Primer (Springer, Berlin, 2015)
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2021 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Jo, T. (2021). Ensemble Learning. In: Machine Learning Foundations. Springer, Cham. https://doi.org/10.1007/978-3-030-65900-4_13
Download citation
DOI: https://doi.org/10.1007/978-3-030-65900-4_13
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-65899-1
Online ISBN: 978-3-030-65900-4
eBook Packages: EngineeringEngineering (R0)