3.7 Multiple Linear Regression
Multiple linear regression extends simple linear regression for p predictors:
Y=β0+β1X1+β2X2+...+βpXp+ϵi - βj is the average effect on Y from Xj holding all other predictors fixed.
Fit is once again choosing the βj that minimizes the RSS.
Example in book shows that although fitting sales against newspaper alone indicated a significant slope (0.055 +- 0.017), when you include radio in a multiple regression, newspaper no longer has any significant effect. (-0.001 +- 0.006)
3.7.1 Important Questions
Is at least one of the predictors X1, X2, … , Xp useful in predicting the response?
F statistic close to 1 when there is no relationship, otherwise greater then 1.
F=(TSS−RSS)/pRSS/(n−p−1)
Do all the predictors help to explain Y , or is only a subset of the predictors useful?
p-values can help identify important predictors, but it is possible to be mislead by this especially with large number of predictors. Variable selection methods include Forward selection, backward selection and mixed. Topic is continued in Chapter 6.
How well does the model fit the data?
R2 still gives proportion of the variance explained, so look for values “close” to 1. Can also look at RSE which is generalized for multiple regression as:
RSE=√1n−p−1RSS
Given a set of predictor values, what response value should we predict, and how accurate is our prediction?
Three sets of uncertainty in predictions:
- Uncertainty in the estimates of βi
- Model bias
- Irreducible error ϵ