Chapter 6 Linear Model Selection and Regularization
Learning objectives:
- Select a subset of features to include in a linear model.
- Compare and contrast the forward stepwise, backward stepwise, hybrid, and best subset methods of subset selection.
- Use shrinkage methods to constrain the flexibility of linear models.
- Compare and contrast the lasso and ridge regression methods of shrinkage.
- Reduce the dimensionality of the data for a linear model.
- Compare and contrast the PCR and PLS methods of dimension reduction.
- Explain the challenges that may occur when fitting linear models to high-dimensional data.
Context for This Chapter
lm(y ~ ., data)
Why constrain or remove predictors?
- improve prediction accuracy
- low bias (by assumption)
- … but \(p \approx n\) -> high variance
- … or meaninglessness \(p = n\)
- … or impossibility \(p > n\)
- model interpretability
- remove or constrain irrelevant variables to simplify model.