![]() | This article may be too technical for most readers to understand.(September 2023) |
Part of a series on |
Machine learning and data mining |
---|
In machine learning (ML), boosting is an ensemble learning method that combines a set of less accurate models (called "weak learners") to create a single, highly accurate model (a "strong learner"). Unlike other ensemble methods that build models in parallel (such as bagging), boosting algorithms build models sequentially. Each new model in the sequence is trained to correct the errors made by its predecessors. This iterative process allows the overall model to improve its accuracy, particularly by reducing bias.[1] Boosting is a popular and effective technique used in supervised learning for both classification and regression tasks.[2]
The theoretical foundation for boosting came from a question posed by Kearns and Valiant (1988, 1989):[3][4] "Can a set of weak learners create a single strong learner?" A weak learner is defined as a classifier that performs only slightly better than random guessing, whereas a strong learner is a classifier that is highly correlated with the true classification. Robert Schapire's affirmative answer to this question in a 1990 paper led to the development of practical boosting algorithms.[5][6] The first such algorithm was developed by Schapire, with Freund and Schapire later developing AdaBoost, which remains a foundational example of boosting.[7]
Arcing [Boosting] is more successful than bagging in variance reduction
The term boosting refers to a family of algorithms that are able to convert weak learners to strong learners
Schapire (1990) proved that boosting is possible. (Page 823)