Professional Documents
Culture Documents
Linear Regression Analysis 6th Edition Montgomery, Peck & Vining 1
Linear Regression Analysis 6th Edition Montgomery, Peck & Vining 1
Chapter 1
Introduction
Designed Experiment
Chapter 2
Simple Linear Regression
( ) ( )
Var y | x = Var β0 + β1x + ε = σ 2
for all levels of x
S ( β 0 , β1=
) ∑i
ε
i
2
= ∑ i 0 1i
( y
i
− β − β x ) 2
What is the LS estimator of μ in univariate statistics?
Answer: Xbar
∂S
−2∑ ( yi − βˆ0 − βˆ1 xi )xi =
= 0
∂β1 βˆ0 , βˆ1 i
i i
LS estimators and also the Maximum Likelihood
n
i =1 i =1
estimators.
i =1 n i =1
∑ n
y ∑
n
x
i i
n
Sxy = ∑ y i x i − i =1 i =1 n
= ∑ y i (x i − x )
i =1 n i =1
Linear Regression Analysis 6E Montgomery, Peck 8
& Vining
Page 31
• Then
∑ n
n
y i ∑ x i
n
y x − i =1 i =1
∑ i i S
ˆβ = i =1 n =
xy
1 2
∑ n
Sxx
xi
n
x 2
− i =1
∑ i
i =1 n
• Residuals:
e i = y i − ŷ i
observed - fit
data - fit
When p is low,
H0 must go.
S xx i =1
Useful in showing expected
value and variance properties
where
n n
ci = ( xi − x ) / S xx , ∑ ci = 0, ∑ ci2 = 1
i =1 i =1
n S
( )
xx
2.2.3 Estimation of σ2
= SST − βˆ1S xy
2.2.3 Estimation of σ2
• Unbiased estimator of σ2
SS Re s
σˆ
= 2
= MS Re s =s2
n−2
s is known as the estimated standard error of the regression model
2.2.3 Estimation of σ2
2
• σ̂ depends on the residual sum of squares. Then:
– Any violation of the assumptions on the model errors could
damage the usefulness of this estimate
– A misspecification of the model can damage the usefulness of
this estimate
– This estimate is model dependent
H0: β1 = 0 H1: β1 ≠ 0
• This tests the significance of regression; that is, is there a
linear relationship between the response and the regressor?
• Failing to reject β1 = 0, implies that there is no linear
relationship between y and x. Does X help to predict Y?
Under H : E(Y) = β 0 0
Y |
|
|
β0 ----------------------------
|
|
-----------------------------
X
Testing
significance
of regression
∑ ( yi − y ) = ∑ ( yˆi − y ) + ∑ ( yi − yˆi )
2 2 2
∑ ( yi − y ) = ∑ ( yˆi − y ) + ∑ ( yi − yˆi )
2 2 2
n −1 = 1 + (n − 2)
• Mean Squares
SS R SS Re s
=MS R = MS Re s
1 n−2
t02 = F0
So for testing significance of regression, the t-test and the
ANOVA procedure are equivalent (only true in simple linear
regression)
( )
βˆ1 − tα / 2,n − 2 se βˆ1 ≤ β1 ≤ βˆ1 + tα / 2,n − 2 se βˆ1 ( )
• 100(1-α)% Confidence interval for the Intercept
( )
βˆ0 − tα / 2,n − 2 se βˆ0 ≤ β 0 ≤ βˆ0 + tα / 2,n − 2 se βˆ0 ( )
• Let x0 be the level of the regressor variable at which we want to estimate the
mean response, i.e.
E ( y | x0 ) = µ y| x0
• The variance of µˆ y| x0 is
( ) ( )
ˆ y| x0 = Var βˆ 0 + βˆ 1 x0 = Var[ y + βˆ 1 ( x0 − x )]
Var µ
= Var ( y ) + Var[βˆ 1 ( x0 − x )]
σ 2 σ 2 ( x0 − x ) 2 1 ( x − x ) 2
= + = σ2 + 0
n S xx n S xx
1 ( x0 − x ) 2
µˆ y| x0 − t α / 2,n − 2 MS Re s + ≤ E ( y | x0 )
n S xx
1 ( x0 − x ) 2
≤ µˆ y| x0 + t α / 2,n − 2 MS Re s +
n S xx
Notice that the length of the CI depends on the location of the point of interest
ŷ= ˆ + βˆ x
β
0 0 1 0
1 ( x0 − x ) 2
≤ yˆ 0 + t α / 2, n − 2 MS Re s 1 + +
n S xx
• R2 can be misleading!
– Simply adding more terms to the model will increase R2
– As the range of the regressor variable increases (decreases), R2
generally increases (decreases).
– R2 does not indicate the appropriateness of a linear model
Identical to OLS
Biased
Also,