Boosting refers to a family of algorithms which converts weak learners to strong learners. How does it happen?
3 Answers
As @desertnaut mentioned in the comment
No weak learner becomes strong; it is the ensemble of the weak learners that turns out to be strong
Boosting is an ensemble method that integrates multiple models(called as weak learners) to produce a supermodel (Strong learner).
Basically boosting is to train weak learners sequentially, each trying to correct its predecessor. For boosting, we need to specify a weak model (e.g. regression, shallow decision trees, etc.), and then we try to improve each weak learner to learn something from the data.
AdaBoost is a boosting algorithm where a decision tree with a single split is used as a weak learner. Also, we have gradient boosting and XG boosting.
- 326
- 2
- 10
You take a bunch of weak learners, each of them trained on a subset of the data.
You then just get all of them to make a prediction, and you learn how much you can trust each one, resulting in a weighted vote or other type of combination of the individual predictions.
- 615
- 3
- 9
In Boosting, we improve the overall metrics of the model by sequentially building weak models and then building upon the weak metrics of previous models.
We start out by applying basic non-specific algorithms to the problem, which returns some weak prediction functions by taking arbitrary solutions (like sparse weights or assigning equal weights/attention). We improve upon this in the following predictions by adjusting weights to those having a higher error rate. After going through many iterations, we combine it to create a single Strong Prediction Function which has better metrics.
Some popular Boosting Algorithms :
- AdaBoost
- Gradient Tree Boosting
- XGBoost
- 750
- 1
- 8
- 20
