Ensemble Models
Frank Acito
Additional contact information
Frank Acito: Indiana University
Chapter Chapter 12 in Predictive Analytics with KNIME, 2023, pp 255-265 from Springer
Abstract:
Abstract Ensemble models in machine learning involve combining predictions from multiple diverse models to achieve improved accuracy and stability. This chapter explores various ensemble techniques and their benefits. The search for the best machine learning algorithm for a particular problem is an ongoing challenge. Studies have shown that no single algorithm performs best across all datasets. This has led to the concept of ensemble learning, where the predictions of multiple models are aggregated to produce a final estimate. The effectiveness of combining diverse independent estimates was first highlighted in “The Wisdom of Crowds.” A classic example by Sir Francis Galton demonstrated the power of combining individual estimates, leading to a more accurate prediction. Ensemble models are created using different approaches, such as employing multiple algorithms, varying model parameters, sampling different subsets of predictor variables, or sampling observations. The benefits of ensemble models lie in reduced variation and improved accuracy. Reduced variation ensures reliability in predictions with different data samples, allowing for a better understanding of the model’s performance with unseen data. Improved accuracy is achieved by combining independent predictions, which helps cancel out errors, resulting in better overall predictions. Bagging, Random Forests, AdaBoost, Gradient Tree Boosting, and XGBoost are discussed. These models are popular for their ability to handle different types of data and achieve state-of-the-art performance in various contexts. The chapter includes practical examples of ensemble modeling with continuous and binary targets. One example uses a KNIME workflow to predict used car prices using ordinary least regression (OLS) and Gradient Boosted Trees. Another example involves predicting credit status using XGBoost.
Date: 2023
References: Add references at CitEc
Citations:
There are no downloads for this item, see the EconPapers FAQ for hints about obtaining it.
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:spr:sprchp:978-3-031-45630-5_12
Ordering information: This item can be ordered from
http://www.springer.com/9783031456305
DOI: 10.1007/978-3-031-45630-5_12
Access Statistics for this chapter
More chapters in Springer Books from Springer
Bibliographic data for series maintained by Sonal Shukla () and Springer Nature Abstracting and Indexing ().