Chapter 10 Resampling for evaluating performance
Learning objectives:
- Recognize why naive performance estimates can often fail.
- Explain the difference between low bias models and high bias models.
- Use resampling to divide a training set into an analysis set and an assessment set.
- Use cross-validation to resample a training set.
- Compare repeated cross-validation, leave-one-out cross-validation, and Monte Carlo cross-validation.
- Divide a “not testing” set into a single training set and a single validation set.
- Use bootstrap resampling to divide a training set into an analysis set and an assessment set.
- Use rolling forecast origin resampling to divide a training set into an analysis set and an assessment set.
- Use resampling to estimate model performance.
- Use
tune::fit_resamples()
to fit multiple models for a resampled dataset. - Use
tune::collect_metrics()
to measure model performance. - Use
tune::collect_predictions()
to analyze model predictions.
- Use
- Use parallel processing to speed up resample fitting.
- Save model objects created during resampling.