IMBoost: A New Weighting Factor for Boosting to Improve the Classification Performance of Imbalanced Data
SeyedEhsan Roshan,
Jafar Tanha,
Farzad Hallaji,
Mohammad-reza Ghanbari and
Giacomo Fiumara
Complexity, 2023, vol. 2023, 1-19
Abstract:
Imbalanced datasets pose significant challenges in the field of machine learning, as they consist of samples where one class (majority) dominates over the other class (minority). Although AdaBoost is a popular ensemble method known for its good performance in addressing various problems, it fails when dealing with imbalanced data sets due to its bias towards the majority class samples. In this study, we propose a novel weighting factor to enhance the performance of AdaBoost (called IMBoost). Our approach involves computing weights for both minority and majority class samples based on the performance of classifier on each class individually. Subsequently, we resample the data sets according to these new weights. To evaluate the effectiveness of our method, we compare it with six well-known ensemble methods on 30 imbalanced data sets and 4 synthetic data sets using ROC, precision-eecall AUC, and G-mean metrics. The results demonstrate the superiority of IMBoost. To further analyze the performance, we employ statistical tests, which confirm the excellence of our method.
Date: 2023
References: Add references at CitEc
Citations:
Downloads: (external link)
http://downloads.hindawi.com/journals/complexity/2023/2176891.pdf (application/pdf)
http://downloads.hindawi.com/journals/complexity/2023/2176891.xml (application/xml)
Related works:
This item may be available elsewhere in EconPapers: Search for items with the same title.
Export reference: BibTeX
RIS (EndNote, ProCite, RefMan)
HTML/Text
Persistent link: https://EconPapers.repec.org/RePEc:hin:complx:2176891
DOI: 10.1155/2023/2176891
Access Statistics for this article
More articles in Complexity from Hindawi
Bibliographic data for series maintained by Mohamed Abdelhakeem ().