Multiple Regression I 4/9/12 Transformations The model Individual coefficients R 2 ANOVA for...

44
Multiple Regression I 4/9/12 • Transformations • The model • Individual coefficients • R 2 • ANOVA for regression • Residual standard error Section 9.4, 9.5 Professor Kari Lock Morgan Duke University

Transcript of Multiple Regression I 4/9/12 Transformations The model Individual coefficients R 2 ANOVA for...

Multiple Regression I4/9/12

• Transformations • The model• Individual coefficients• R2 • ANOVA for regression• Residual standard error

Section 9.4, 9.5 Professor Kari Lock MorganDuke University

• Project 2 Proposal (due Wednesday, 4/11)

• Homework 9 (due Monday, 4/16)

• Project 2 Presentation (Thursday, 4/19)

• Project 2 Paper (Wednesday, 4/25)

To Do

Non-Constant Variability

Non-Normal Residuals

Transformations• If the conditions are not satisfied, there are some common transformations you can apply to the response variable

• You can take any function of y and use it as the response, but the most common are• log(y) (natural logarithm - ln)• y (square root)• y2 (squared)• ey (exponential))

log(y)Original Response, y:

Logged Response, log(y):

yOriginal Response, y:

Square root of Response, y:

y2

Original Response, y:

Squared response, y2:

ey

Original Response, y:

Exponentiated Response, ey:

• Multiple regression extends simple linear regression to include multiple explanatory variables:

Multiple Regression

0 1 2 21 ... k k ix xxy ò

• We’ll use your current grades to predict final exam scores, based on a model from last semester’s students

• Response: final exam score

• Explanatory: hw average, clicker average, exam 1, exam 2

Grade on Final

0 1 2 3 4hw clicker exam1 exam2y

What variable is the most significant predictor of final exam score?

a) Homework averageb) Clicker averagec) Exam 1 d) Exam 2

Grade on Final

The p-value for explanatory variable xi is associated with the hypotheses

For intervals and p-values of coefficients in multiple regression, use a t-distribution with degrees of freedom n – k – 1, where k is the number of explanatory variables included in the model

0 : 0

: 0a

i

iH

H

Inference for Coefficients

Estimate your score on the final exam.

What type of interval do you want for this estimate?

a) Confidence intervalb) Prediction interval

Grade on Final

Estimate your score on the final exam.(hw average is out of 10, clicker average is out of 2)

Grade on Final

Is the clicker coefficient really negative?!?

Give a 95% confidence interval for the clicker coefficient (okay to use t* = 2).

Grade on Final

Is your score on exam 2 really not a significant predictor of your final exam score?!?

Grade on Final

• The coefficient (and significance) for each explanatory variable depend on the other variables in the model!

• In predicting final exam scores, if you know someone’s score on Exam 1, it doesn’t provide much additional information to know their score on Exam 2 (both of these explanatory variables are highly correlated)

Coefficients

If you take Exam 1 out of the model…

Grade on Final

Model with Exam 1:

Now Exam 2 is significant!

If you include Project 1 in the model…

Grade on Final

Model without Project 1:

Grades

Multiple Regression• The coefficient for each explanatory variable is the

predicted change in y for one unit change in x, given the other explanatory variables in the model!

• The p-value for each coefficient indicates whether it is a significant predictor of y, given the other explanatory variables in the model!

• If explanatory variables are associated with each other, coefficients and p-values will change depending on what else is included in the model

Residuals

Are the conditions satisfied?(a) Yes (b) No

Evaluating a Model

• How do we evaluate the success of a model?

• How we determine the overall significance of a model?

• How do we choose between two competing models?

Variability• One way to evaluate a model is to partition variability

• A good model “explains” a lot of the variability in Y

Total Variability

VariabilityExplained

by the Model

Error Variability

Exam Scores• Without knowing the explanatory variables, we can say that a person’s final exam score will probably be between 60 and 98 (the range of Y)

• Knowing hw average, clicker average, exam 1 and 2 grades, and project 1 grades, we can give a narrower prediction interval for final exam score

• We say the some of the variability in y is explained by the explanatory variables

• How do we quantify this?

VariabilityHow do we quantify variability in Y?

a) Standard deviation of Yb) Sum of squared deviations from the

mean of Yc) (a) or (b)d) None of the above

Sums of Squares

2

1

n

ii

Y Y

Total Variability

VariabilityExplained

by the model

Error variability

2

1

ˆn

ii

Y Y

2

1

ˆn

i ii

Y Y

SST SSM SSE

Variability

Y

2

1

Total Sum of Squares:n

ii

SST y y

2

1

Model Sum of Squares:

ˆn

ii

SSM y y

2

1

Error Sum of Squares:

ˆn

i ii

SSE y y

• If SSM is much higher than SSE, than the model explains a lot of the variability in Y

R2

• R2 is the proportion of the variability in Y that is explained by the model

2 "Variability in Y explained by the model"

"Total variability in Y"

SSMR

SST

Total Variability

Variability Explained by the Model

R2

• For simple linear regression, R2 is just the squared correlation between X and Y

• For multiple regression, R2 is the squared correlation between the actual values and the predicted values

R2

2 0.67R 2 0.09R

Final Exam Grade

Is the model significant?• If we want to test whether the model is significant (whether the model helps to predict y), we can test the hypotheses:

• We do this with ANOVA!

0 1 2: ... 0

: At least one 0k

a i

H

H

ANOVA for Regression

k: number of explanatory variablesn: sample size

Source

Model

Error

Total

df

k

n-k-1

n-1

Sum ofSquares

SSM

SSE

SST

MeanSquare

MSM = SSM/k

MSE = SSE/(n-k-1)

F

MSMMSE

p-value

Use Fk,n-k-1

Final Exam GradeFor this model, do the explanatory variables significantly help to predict final exam score? (calculate a p-value).

(a) Yes

(b) No

n = 69SSM = 3125.8SSE = 1901.4

ANOVA for Regression

Source

Model

Error

Total

df

5

63

68

Sum ofSquares

3125.8

1901.4

5027.2

MeanSquare

625.16

30.18

F

20.71

p-value

0

Final Exam Grade

Simple Linear Regression• For simple linear regression, the following tests will all give equivalent p-values:

• t-test for non-zero correlation

• t-test for non-zero slope

• ANOVA for regression

Mean Square Error (MSE)• Mean square error (MSE) measures the average variability in the errors (residuals)

• The square root of MSE gives the standard deviation of the residuals (giving a typical distance of points from the line)

• This number is also given in the R output as the residual standard error, and is known as s in the textbook

Final Exam Grade

0 1i i iy x

Simple Linear Model

~ 0,i N

Residual standard error = MSE = se estimates the standard deviation of

the residuals (the spread of the normal distributions around the

predicted values)

Residual Standard Error• Use the fact that the residual standard error is 5.494 and your predicted final exam score to compute an approximate 95% prediction interval for your final exam score

• NOTE: This calculation only takes into account errors around the line, not uncertainty in the line itself, so your true prediction interval will be slightly wider

ˆ 2 5.494y

• How do we decide which explanatory variables to include in the model?

• How do we use categorical explanatory variables?

• What is the coefficient of one explanatory variable depends on the value of another explanatory variable?

To Come…