Chapter 10 Bagging (Bootstrap Aggregating)
It’s a great way to the reduce model variance \(\text{Var}(\hat{f}(x_0))\) and improve accuracy by reducing overfitting.
This method consist in:
- Creating bootstrap copies of the original training data
- Fitting multiple (\([50,500]\)) versions of a base learner (model), as we need fewer resamples if we have strong predictors
- Combining models into an aggregated prediction
- Regression: By averaging the predictions.
- Classification: By averaging the estimated class probabilities or using the plurality vote.