Voting and Bagging
Christo El Morr,
Manar Jammal,
Hossam Ali-Hassan and
Walid El-Hallak ()
Additional contact information
Christo El Morr: York University
Manar Jammal: York University
Hossam Ali-Hassan: York University, Glendon Campus
Walid El-Hallak: Ontario Health
Chapter Chapter 14 in Machine Learning for Practical Decision Making, 2022, pp 413-430 from Springer
Abstract:
Abstract The ensemble technique relies on the idea that aggregation of many classifiers and regressors will lead to a better prediction [1]. In this chapter, we will introduce the ensemble technique and cover two ways in which to organize an ensemble (literally, a set) of machine learning methods called voting and bagging [2] and one algorithm to perform bagging called random forest [1, 3]. The other two ways to organize the ensemble methods are called boosting and stacking, which will be covered in the next chapter.
Date: 2022
References: Add references at CitEc
Citations:
There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:isochp:978-3-031-16990-8_14
Ordering information: This item can be ordered from
http://www.springer.com/9783031169908
DOI: 10.1007/978-3-031-16990-8_14
Access Statistics for this chapter
More chapters in International Series in Operations Research & Management Science from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().