Out-of-bag (OOB) error

As we already have resample the data several times, we don’t to do it again to estimate the test error if we have enough resamples \(B\).

The out-of-bag (OOB) error estimation is approximation to the leave-one-out cross validation by taking advantage of the \(1/3\) observations that weren’t used to train each base learner which opens the possibility to calculate \(B/3\) predictions, aggregate them and calculate the test error.

As your data sets become larger and your bagging iterations increase, it is common to use the OOB error estimate as a proxy for predictive performance.

The authors explained an example estimating the RMSE of 200 trees

Method RMSE Needed Time
10-fold cross validation 26,957 26 minutes
OOB error 26,462 0.96 minutes