Professional Documents
Culture Documents
GLS+ WLS+ Ols
GLS+ WLS+ Ols
Answer: GLS is a method of finding the best estimator of 𝛽 in the model 𝑌=X𝛽+u
when E(u) = 0, and E(u, u՛)𝜎 2 v where v is n×n positive definite matrix.
To find the GLS estimator of 𝛽 the basis Idea is to transform are observations y
to other variables 𝑦*. Which satisfy the usual assumptions and then apply the OLS
method to the variable so obtained.
Answer: One of the fundamental assumptions in regression about the random vector u
is that E(u1 𝑢́ ) =𝜎 2 𝐼 which is known as the assumption of homoscedasticity. If these
assumption is violated i.e. u՛s are heteroscadasticity or auto correlated. Then we apply
GLS to find the best estimator of 𝛽.
5Question:
Which assumption of OLS method when violated necessity. The use
of GLS method for obtaining best estimator.
β̂ = (x′ x)-1x′y
= (x′ x)-1 x′ (xβ+𝑢)
= β+ (x′x)-1 x′ 𝑢
𝑥1
𝑥2
. 2
Now x′ x = [𝑥1 𝑥2 . . . 𝑥𝑛 ] . = ∑𝑥𝑖
.
(𝑥𝑛 )
1
(x′x)-1 = and x′ 𝑢 = ∑𝑥𝑖 𝑢𝑖
∑𝑥𝑖2
∑𝑥 𝑢
∴ β̂ = 𝑏̂1 = b1 + ∑𝑥𝑖 2𝑖
𝑖
E (β̂) = b1
In matrix notation
Y = 𝑋𝛽+𝑈
𝑦1 1 𝑥1 𝑢1
𝑦2 1 𝑥2 𝑢2
. . . 𝑏 .
Where 𝑦 = . x= 𝛽 = ( 0) 𝑢= .
. . 𝑏1
. . . .
(𝑦𝑛 ) (1 𝑥𝑛 ) (𝑢𝑛 )
0 𝑥2
1
. . . 0 𝑦2
1 1 . . . 1 .2 .
x′v-1y = [ ] . . . . .
.
𝑥1 𝑥2 0 0 0 𝑥𝑛 . . . . . .
. . . . . . .
1 (𝑦𝑛 )
(0 0 . . . 2)
𝑥𝑛
𝑦1
1 1 1 𝑦2 𝑦
. . . 𝑥2 . ∑ 𝑥 2𝑖
𝑥12 𝑥22 𝑛 𝑖
= [1 1 1] . = [ 𝑦𝑖 ]
0 0 0 ∑
𝑥1 𝑥2 𝑥2 . 𝑥𝑖
(𝑦𝑛 )
1 𝑦
𝑛 − 𝛴𝑥 ∑ 𝑥 2𝑖
1
β̃= (x′ v-1 x)-1 x′ v-1y = 2 [ 1 1
𝑖
][ 𝑖
𝑦𝑖 ]
1 1
𝑛∑ 2 − (∑𝑥 ) − 𝛴𝑥 ∑𝑥
𝑥 𝑖 𝑖 𝑖 ∑𝑥𝑖2 𝑖
𝑦𝑖 1 𝑦𝑖
𝑛∑ 𝑥 2 − ∑
1 𝛴𝑥𝑖 𝑥𝑖
β̃ = 2 [ 1
𝑖
𝑦 1 𝑦𝑖 ]
𝑛∑
1
𝑥2
1
– (∑𝑥 ) − 𝛴𝑥 ∑ 𝑥𝑖 + ∑𝑥
𝑖 𝑖 𝑖 𝑖 ∑𝑥𝑖2 𝑖
1 𝑦 1 𝑦
∑𝑥𝑖 − ∑𝑥 ∑𝑥𝑖
∑𝑥2 𝑖 𝑖 𝑖
𝑏̃1 = 𝑖
2
1 1
𝑛∑ − (∑𝑥 )
𝑥2
𝑖 𝑖
1
𝑛 −∑ 𝑥
𝑘2
𝜈 (β̃) = 𝜈 (𝑏̃1 ) = 𝑘 2 (x′ v-1 x)-1 = 𝑖
1 1
2 [ 1 1 ]
𝑛∑ 2 − (∑𝑥 ) − 𝛴𝑥
𝑥𝑖 𝑖 𝑖 ∑𝑥𝑖2
1
𝑘2∑
𝑥2
𝜈 (𝑏̃1 ) = 𝑖
2 (ii)
1 1
𝑛∑ − (∑𝑥 )
𝑥2
𝑖 𝑖
0
1
. . . 0 𝑥2
𝑥22 .
∴ (x′ v-1 x) = [x1 x2 . . . xn ] . . . . . .
.
. . . . . .
. . . . . . .
1 (𝑥𝑛 )
(0 0 . . . 2)
𝑥𝑛
𝑥1
𝑥2
1 1 1 .
= [x x . . . x ] .
1 2 n
.
(𝑥𝑛 )
= [1 + 1 + 1+. . . +1] = n
8Question: State and proof Gass –Aitkens theorem on GLS estimator OR Show
that GLSE is BLUE.
Statement: Let in general linear regression or general linear regression model of full
rank Y = × β + u, E(u) = 0, 𝜈 (u) = 𝜎 2 v
Where 𝜈 is positive definite matrix. Then the GLS estimator of 𝛽 is the best linear
unbiased estimator i.e. GLSE is BLUE
Unbiasedness: We have
∴ E(β*) = β + M × β + 0 [⸪ A × = I]
β = β+M× β
M× β = 0 [⸪ β ≠ = 0]
M× = 0
=> x′ M′ = 0
Now β* = A× β + M× β + (A+M) u
= β+ ∗ β + (A+M) u
= β + (A + M) u
*
=> β - β = (A + M) u
1 0 0 . . 0
1 1
0 2 0 . . 0
2
1
0 0 3 . . . .
x′ 𝜈 -1 x = [1 2 . . .n]
. . . . . . .
. . . . . . .
1 (𝑛)
(0 0 0 . . 𝑛 )
1
2
.
= [1 1 . . . 1] = 1+2+3+ . . . + n
.
.
(𝑛)
𝑛(𝑛+1)
= 2
𝑦1
𝑦2
2 .
(x′ 𝜈 -1x)-1 = and x′ 𝜈 -1 𝛾 = [1 1 . . . 1] . => ∑𝑦𝑖
𝑛(𝑛+1)
.
(𝑦𝑛 )
2∑𝑦
𝑖
𝜃* = (x′ 𝜈 -1x)-1 x′ 𝜈 -1 y = 𝑛(𝑛+1)
2𝑘
𝜈 (𝜃*) = k (x′ 𝜈 -1x)-1 = 𝑛(𝑛+1)
1
2
.
= [12 22 . . . n2] = 13 + 23 +33 ------+ n3
.
.
(𝑛)
𝑛(𝑛+1) 2
={ }
2
𝑛(𝑛+1) 2 6 2
𝜈 ( 𝜃̂ ) = k { } { }
2 𝑛(𝑛+1)(2𝑛+1)
9𝑘 9 2𝑘 9 2𝑘
= (2𝑛+1)2
= 2 ≤
2 (2𝑛+1) 2 𝑛(𝑛+1)
=> 𝜈 ( 𝜃̂ ) ≤ 9⁄2 𝜈 (𝜃*)
=> 𝜈 (𝜃*) < 𝜈 ( 𝜃̂ )
Hence GLS is more efficient than OLS.
Question:
10 For the model Yi = βxi + Ui, Ui ~ NID (0, i𝜎 2 ) i= 1, 2 if x1 = 1
And x2 = -1 obtain the GLS of β and find its variance.
In matrix notation
y1 1 u1
(y ) = ( ) 𝛽 + (u )
2 −1 2
Y = Xβ + U
Now Ui ~ NID (0, i𝜎 2 )
2
∴ E(u) = 0 and E(u u՛) = i 𝜎 2 = ( 𝜎 0 ) = 𝜎 2 (1 0)
0 2𝜎 2 0 2
1 0 -1 1 0
Where 𝜈= ( ) 𝜈 = (0 1 )
0 2 ⁄2
Thus the GLS of β is
β̂ = (x′𝜈 -1x)-1 x′𝜈 -1y
1 0 1 1 1
(x′ 𝜈 -1x) = (1 -1) (0 1 ) ( ) = (1 − ⁄2 ) ( ) = 1+1⁄2 = 3⁄2
⁄2 −1 −1
1 0 y1 1 y1 y
x′ 𝜈 -1 𝛾 = (1 -1) (0 1 ) (y ) = (1 − ⁄2 ) (y ) = y1 - 2⁄2
⁄2 2 2
γ2
∴ β̂ = 3 (y1 − ⁄2) = 2⁄3 y1 - 32
2 γ
11 Question: Let y𝑖 ~ NID (0, 𝜎 2⁄𝜔𝑖 ) ; i = 1,2 … n Find the linear unbiased
estimator of 𝜃 with minimum variance.
Answer: Since y𝑖 ~ NID (0, 𝜎 2⁄𝜔𝑖 ) ; i = 1,2 … n we get the linear model y𝑖 = 𝜃+ui;
2
i = 1,2, …, n and E (ui) = 0, E(𝑢𝑖 2 ) = 𝜎 ⁄𝑤𝑖 and E (ui uj) = 0 ∀ i ≠ j In matrix notation
we can write this model as
Y = Xβ+U
𝑦1 1 𝑢1
𝑦2 1 𝑢2
. . .
where Y = . ,𝑥= , and u = .
.
. . .
(𝑦𝑛 ) (1) (𝑢𝑛 )
𝜎 2⁄ 0 . . . 0
𝜔1
0 𝜎 2⁄ . . . 0
𝜔2
E (u) = 0, E(u, u′) = . . . . . . 𝜎 2 𝜈 (say)
. . . . . .
. . . . . .
2
( 0 0 . . . 𝜎 ⁄𝜔𝑛 )
1⁄ 0 . . . 0
𝜔1
0 1⁄ . . . 0
𝜔2
where 𝜈 = . . . . . .
. . . . . .
. . . . . .
( 0 0 . . . 1⁄𝜔𝑛 )
𝜔1 0 . . . 0
0 𝜔2 . . . 0
. . . . . .
𝜈 -1 =
. . . . . .
. . . . . .
(0 0 . . . 𝜔𝑛 )
The GLS of 𝜃 is
𝜃̂ = (x′ 𝜈 -1x)-1 x′ 𝜈 -1 y
𝜔1 0 . . . 0 1
0 𝜔2 . . . 0 1
. . . . . . .
x′ 𝜈 -1x = [1 1 … 1]
. . . . . . .
. . . . . . .
(0 0 . . . 𝜔𝑛 ) 1)
(
1
1
.
= [w1 w2 … wn ] = ∑ 𝜔𝑖
.
.
(1)
𝜔1 0 . . . 0 𝑦1
0 𝜔2 . . . 0 𝑦2
. . . . . . .
x′ 𝜈 -1 y= [1 1 …. 1] . = ∑ 𝜔𝑖 𝑦𝑖
. . . . . .
. . . . . . .
(0 0 . . . 𝜔𝑛 ) (𝑦𝑛 )
∴ 𝜈 ( 𝜃̂ ) = 𝜎 2 (x′ 𝜈 -1x)-1
1 2 ∑𝜔 𝑦
= 𝜎 2 . ∑ 𝜔 = 𝜎 ⁄∑ 𝜔 and 𝜃̂ ∑ 𝜔𝑖 𝑖 with minimum variance.
𝑖 𝑖 𝑖
Y=Xβ+U . . . (1)
(y-xb)՛v-1 (y-xb)
= (y՛-b՛x՛)v-1 (y-xb)
2bx՛v-1x = 2 x՛v-1y
b = (x՛ v-1x)-1 x՛ v-1y
The estimator of b is b = = (x՛ v-1x)-1 x՛ v-1y
=β+0
=β
=b
= E[(𝑏̂ − 𝑏)( 𝑏̂ − 𝑏) ՛]
= σ2(x՛ v-1x)-1
Answer: The OLSE of β is 𝛽̂ =(x՛x)-1 x՛y. Another application for this problem by
transforming the model is a n set of observation that satisfy the Standard Least Square
assumption. Then we shall use OLS on the transformation data. Since σ2v is the
dispersion matrix of the errors v must be positive definite. So there exists an n×n no
singular matrix p such that
PPT = V
Let P be a transformation matrix such that
Y=Xβ+U
TY=TXβ+TU
Let TY = Z, TX = W, TU = C
Z = Wβ + C . . . (2)
= T σ2 V T՛
= σ TV
2
T՛
Then we have E( CC՛) = σ2 I and we can apply OLS method to the model … (i)
Now PP՛ = V
P-1 V (P՛)-1 = 1
Hence the appropriate T = P-1
β* = (W՛W) -1 W՛Z
= (PP՛)-1
= V-1
Yi = 𝛽1 + 𝛽2 𝑥𝑖 + 𝑢𝑖 ; ii= 1,2, … n
1
0 . . . 0
𝑥1
1
0 . . . 0
2 𝑥2
With E(u) = 0, E (uiuj) = 𝜎 . . . . . . = 𝜎 2 𝑣 (say)
. . . . . .
. . . . . .
(0 0 . . . 𝑥𝑛 )
1
0 . . . 0
𝑥1 𝑥1 0 0 . . 0
1 0 𝑥2 0 . . 0
0 . . . 0
𝑥2
0 0 𝑥3 . . .
Where v = . . . . . . 𝑣 −1 =
. . . . . . . . . . . .
. . . . . . . . . . . .
1 (0 0 0 . . 𝑥𝑛 )
(0 0 . . . 𝑥𝑛 )
𝛽̂ = (𝑥 ′ 𝑣 −1 𝑥)−1 𝑥 ′ 𝑣 −1 𝑦
𝑥1 0 0 . . 0 1 𝑥1
0 𝑥2 0 . . 0 1 𝑥2
1 1 . . . 1 0 0 𝑥3 . . . . .
(𝑥 ′ 𝑣 −1 𝑥) = ( )
𝑥1 𝑥2 . . . 𝑥𝑛 . . . . . . . .
. . . . . . . .
(0 0 0 . . 𝑥𝑛 ) (1 𝑥𝑛 )
∑𝑥 ∑𝑥𝑖2
= ( 2𝑖 )
∑𝑥𝑖 ∑𝑥𝑖3
1 ∑𝑥𝑖3 −∑𝑥𝑖2
⸫ (𝑥 ′ 𝑣 −1 𝑥)−1 = 2 ( )
𝛴𝑥𝑖 ∑𝑥𝑖3 −(∑𝑥𝑖2 ) −∑𝑥𝑖2 ∑𝑥𝑖
𝑥1 0 0 . . 0 𝑦1
0 𝑥2 0 . . 0 𝑦2
1 1 . . . 1 0 0 𝑥3 . . . .
⸫ (𝑥 ′ 𝑣 −1 𝑦) = ( ) .
𝑥1 𝑥2 . . . 𝑥𝑛 . . . . . .
. . . . . . .
(0 0 0 . . 𝑥𝑛 ) (𝑦𝑛 )
∑𝑥𝑖 𝑦𝑖
= ( 2 )
∑𝑥𝑖 𝑦𝑖
3
1 ∑𝑥𝑖 −∑𝑥𝑖2 ∑𝑥𝑖 𝑦𝑖
𝛽̂ = 2 ( )( 2 )
𝛴𝑥𝑖 ∑𝑥𝑖3 −(∑𝑥𝑖2) −∑𝑥𝑖2 ∑𝑥𝑖 ∑𝑥𝑖 𝑦𝑖
V(𝛽̂ ) = 𝜎 2 (𝑥 ′ 𝑣 −1 𝑥)−1
𝜎2 ∑𝑥𝑖3 −∑𝑥𝑖2
= 2 ( )
𝛴𝑥𝑖 ∑𝑥𝑖3 −(∑𝑥𝑖2 ) −∑𝑥𝑖2 ∑𝑥𝑖
𝜎2 ∑𝑥𝑖3
⸫ V(𝛽̂1 ) = 2 which is the BLUE of 𝛽1
𝛴𝑥𝑖 ∑𝑥𝑖3 −(∑𝑥𝑖2 )
𝜎2 𝛴𝑥𝑖
V(𝛽̂2 ) = 2 which is the BLUE of 𝛽2
𝛴𝑥𝑖 ∑𝑥𝑖3 −(∑𝑥𝑖2 )
Answer: Let (xi, yi) i = 1, 2, … , n be the nth pairs of observations that satisfy the
following variable regression model.
Yi = α+ βXi+ u ; i = 1, 2, … , n
Here the value of α and β that minimize the sum of squares of residuals ∑𝑢𝑖2 are
defined to be the least square estimators of α and β. So the method which minimize the
sum of squares of residuals ∑𝑢𝑖2 called the ordinary least squares method or principle
least square method.
Answer: Regression analysis deals with the dependent of one variable on other
variables but it does not necessarily imply causation.
Answer: The distinguish between linear and non-linear r regression model are as
follows
Linear No – linear
(i) A model which is linear in (i) A model which is not linear
parameter is called linear in parameter is called non
model. Example: linear model. Example: 𝑌 =
𝑌 = 𝛽0 + 𝛽1 𝑥1𝑖 + 𝛽2 𝑥2𝑖 + 𝑋𝛽 + 𝑈
⋯ + 𝛽𝑘 𝑥𝑘𝑖 + 𝑢𝑖 ;
(ii) The model can be fitted by (ii) The model can not be fitted
OLS method. by OLS method.
(iii) The parameter of such model (iv) The parameter of such model
can easily estimated. can not easily estimated.
Then we have
𝑦𝑖 = 𝑦̂𝑖 + 𝑒𝑖 (2)
= ∑𝑦̂𝑖2 + ∑𝑒𝑖2
19
QUESTION: Consider the following formulation of the two variable
population regression function.
Model I: yi = 𝛽1 + 𝛽2 𝑥𝑖 + 𝑢𝑖
yi = 𝛽1 + 𝛽2 𝑥𝑖 + 𝑢𝑖 … (i)
𝑦𝑖 = 𝛼1 + 𝛼2 (𝑥𝑖 − 𝑥̅ ) + 𝑢𝑖 … (ii)
For model-I: Let 𝛽1 𝑎𝑛𝑑 𝛽2 be the OLS estimator of 𝛽1 𝑎𝑛𝑑 𝛽2 respectively. Then the
model becomes
𝑦𝑖 = 𝛽̂1 + 𝛽̂2 𝑥𝑖 + 𝑒𝑖
𝑛 2
∑𝑒𝑖2 = ∑ (𝑦𝑖 − 𝛽̂1 − 𝛽̂2 𝑥𝑖 )
𝑖=1
𝜕∑𝑒𝑖2
=0
𝜕𝛽̂1
𝑛
2 ∑𝑖=1(𝑦𝑖 − 𝛽̂1 − 𝛽̂2 𝑥𝑖 ) (−1) = 0
∑𝑦𝑖 - n𝛽̂1 - 𝛽̂2 ∑𝑥𝑖 = 0
n𝛽̂1 = 2𝑦𝑖 − 𝛽̂2 ∑𝑥𝑖
∑𝑦 ∑𝑥
𝛽̂1 = 𝑛 𝑖 − 𝛽̂2 𝑛 𝑖
𝛽̂1 = 𝑦̅ − 𝛽̂2 𝑥̅ …. (iii)
Again
𝜕∑𝑒𝑖2
=0
𝜕𝛽̂2
𝑛
2 ∑𝑖=1(𝑦𝑖 − 𝛽̂1 − 𝛽̂2 𝑥𝑖 ) (−𝑥𝑖 ) = 0
∑𝑥𝑖 𝑦𝑖 − 𝛽̂1 ∑𝑥𝑖 − 𝛽̂1 ∑𝑥𝑖2 = 0
∑𝑥𝑖2 𝑦𝑖 (∑𝑥 ) 2
∑𝑥𝑖 𝑦𝑖 − + 𝛽̂2 𝑛𝑖 − 𝛽̂2 ∑𝑥𝑖2 = 0
𝑛
(𝛴𝑥𝑖 )2 ∑𝑥𝑖 ∑𝑦𝑖
𝛽̂2 {∑𝑥𝑖2 − } = ∑𝑥𝑖 𝑦𝑖 −
𝑛 𝑛
∑𝑥 ∑𝑦
∑𝑥𝑖 𝑦𝑖 − 𝑖 𝑖
𝛽̂2 = 𝑛
2
(𝛴𝑥 )
∑𝑥𝑖2 − 𝑛𝑖
∑(𝑥𝑖 𝑦𝑖 )−𝑥̅ ∑𝑦𝑖
𝛽̂2 = ∑(𝑥𝑖 −𝑥̅ )2
∑(𝑥𝑖 −𝑥̅ )𝑦𝑖
𝛽̂2 = ∑(𝑥𝑖 −𝑥̅ )2
∑(𝑥𝑖 −𝑥̅ )
𝛽̂2 = 𝑌
∑(𝑥𝑖 −𝑥̅ )2
∑(𝑥 −𝑥̅ )
𝛽̂2 = 𝛴𝑤𝑖 𝑦𝑖 [where ∑𝑤𝑖 = ∑(𝑥 𝑖−𝑥̅ )2 = 0 𝑎𝑛𝑑 ∑𝑤𝑖 𝑥𝑖 = 1]
𝑖
= 𝛽2 + 0
= 𝛽2
∑𝑦𝑖 ∑(𝑥𝑖 − 𝑥̅ )
𝛽̂1 = − 𝑥̅ 𝑦
𝑛 ∑(𝑥𝑖 − 𝑥̅ )2 𝑖
∑𝑦
𝛽̂1 = 𝑖 − 𝑥̅ 𝛴𝑤𝑖̇ 𝑦𝑖
𝑛
1
𝛽̂1 = ∑ ( − 𝑥̅ 𝑤𝑖 ) 𝑦𝑖
𝑛
1
𝛽̂1 = ∑ ( − 𝑥̅ 𝑤𝑖̇ ) 𝑦𝑖
𝑛
1
𝛽̂1 = ∑ (𝑛 − 𝑥̅ 𝑤𝑖̇ ) (𝛽1 + 𝛽2 𝑥𝑖 + 𝑢𝑖 )
1 𝛴𝑥 1
𝛽̂1 = 𝑛 𝑛𝛽1 + 𝛽2 𝑛 𝑖 + ∑ (𝑛 − 𝑥̅ 𝑤𝑖 ) 𝑢𝑖 - 𝛽1 ∑𝑤𝑖 𝑥̅ − 𝑥̅ 𝛽2 𝑤𝑖 𝑥𝑖
1
𝛽̂1 = 𝛽1 + 𝛽2 𝑥̅ + ∑ (𝑛 − 𝑥̅ 𝑤𝑖̇ ) 𝑢𝑖 − 0 − 𝛽2 𝑥̅
1
𝛽̂1 = 𝛽1 + ∑ ( − 𝑥̅ 𝑤𝑖 ) (𝑢𝑖 )
𝑛
1
⸫ 𝐸(𝛽̂1 ) = 𝛽1 + 2 (𝑛 − 𝑥̅ 𝑤𝑖 ) 𝐸(𝑢𝑖 )
= 𝛽1 + 0
= 𝛽1
Now variance
2
𝑣(𝛽̂1 ) = 𝐸(𝛽̂1 − 𝛽1 )
1 2
= 𝐸 [𝛽1 + ∑ (𝑛 − 𝑥̅ 𝑤𝑖 ) 𝑢𝑖 − 𝛽1 ]
1 2
= 𝐸 [∑ ( − 𝑥̅ 𝑤𝑖 ) 𝑢𝑖 ]
𝑛
1 2
= 𝐸 (𝑛 − 𝑥̅ 𝑤𝑖 ) 𝐸(𝑈𝑖2 )
1 1
= 𝜎 2 [∑ (𝑛2 − 2𝑥̅ 𝑛 𝑊𝑖 + 𝑥̅ 2 𝑊𝑖2 ) + 0]
1
= 𝜎 2 [𝑛 𝑛2 − 2𝑥̅𝜋1 ∑𝑤𝑖 + 𝑥̅ 2 𝛴𝑤𝑖2 ]
1 ∑(𝑥 −𝑥̅ )2
= 𝜎 2 [𝑛 − 0 + 𝑥̅ 2 {∑(𝑥 𝑖−𝑥̅ )2 }2 ]
𝑖
1 𝑥̅ 2
= 𝜎 2 [𝑛 + {∑(𝑥 −𝑥̅ )2 }]
𝑖
∑𝑥 2 2𝑛𝑥̅ 2 ∑𝑥𝑖 1
= 𝜎 2 [𝑛∑(𝑥 −𝑥̅
𝑖
)2
+ − 2𝑥̅ ]
𝑖 𝑛∑(𝑥𝑖 −𝑥̅ )2 𝑛 ∑(𝑥𝑖 −𝑥̅ )2
∑𝑥 2 2𝑥̅ 2 2𝑥̅ 2
= 𝜎 2 [𝑛∑(𝑥 −𝑥̅
𝑖
)2
+ ∑(𝑥 −𝑥̅ )2
− ∑(𝑥 −𝑥̅ )2]
𝑖 𝑖 𝑖
∑𝑥 2
= 𝜎 2 [𝑛∑(𝑥 −𝑥̅
𝑖
)2
]
𝑖
𝜎2 ∑𝑥𝑖2
⸫ 𝑣(𝛽̂1 ) = 𝑛∑(𝑥𝑖 −𝑥̅ )2
2
Again 𝑣(𝛽̂2 ) = 𝐸(𝛽̂2 − 𝛽2 )
= E[𝐵2 + ∑𝑤𝑖 𝑢𝑖 − 𝛽2 ]2
= ∑𝑤𝑖2 𝐸(𝑢𝑖2 )
1 ∑(𝑥 −𝑥̅ ) ∑(𝑥 −𝑥̅ )2 1
= 𝜎 2 ∑(𝑥 −𝑥̅ )2 [∑𝑤𝑖̇ = ∑(𝑥 𝑖−𝑥̅ )2 , ∑𝑊𝑖2 = {∑(𝑥 𝑖−𝑥̅ )2 }2 = ∑(𝑥𝑖 −𝑥̅ )2
]
𝑖 𝑖 𝑖
𝑦𝑖 = 𝛼1 + 𝛼2 (𝑥𝑖 − 𝑥̅ ) + 𝑢𝑖
Let 𝛼̂1 𝑎𝑛𝑑 𝛼̂2 be the OLSE of 𝛼1 𝑎𝑛𝑑 𝛼2 respectively then the model becomes
Now differentiating with respect to 𝛼̂1 𝑎𝑛𝑑 𝛼̂2 in (1) and put equal to zero.
𝜕𝛴𝑢𝑖2
̂1
=0
𝜕𝛼
Again
𝜕𝛴𝑢𝑖2
̂2
=0
𝜕𝛼
⸫𝐸[𝛼̂2 ] = 𝛼2
= 𝐸(𝛼2 + ∑𝑤𝑖 𝑢𝑖 − 𝛼2 )2
= 𝐸(𝛴𝑤𝑖 𝑢𝑖 )2
=∑𝑤𝑖2 𝛴(𝑢𝑖2 )
= 𝜎 2 𝛴𝑊𝑖2
𝜎2 ∑(𝑥 −𝑥̅ ) 1
= ∑(𝑥 −𝑥̅ )2 [∑𝑤𝑖 = ∑(𝑥 𝑖−𝑥̅ )2 , ∑𝑤𝑖2 = ∑(𝑥 −𝑥̅ )2]
𝑖 𝑖 𝑖
Now 𝛼̂1 = 𝑦̅
∑𝑦𝑖
𝛼̂1 =
𝑛
1 1 1
= 𝑛 ∑𝛼𝑖 + 𝑛 𝛼1 𝛴(𝑥𝑖 − 𝑥̅ ) + 𝑛 ∑𝑢𝑖
𝑛𝛼1 𝛼2 𝛴𝑢𝑖
= + ×0+
𝛼1 𝑛 𝑛
∑𝑢𝑖
𝑎̂1 = 𝛼1 +
𝑛
1
𝐸[𝑎̂1 ] = 𝛼1 + ∑𝐸(𝑢𝑖 )
𝑛
= 𝛼1 + 0
= 𝛼1
Comment:
(a) 𝛽1 𝑎𝑛𝑑 𝛼1 are not identical. Because 𝛼̂1 = 𝑦̅ and 𝛽̂1 = 𝑦̅ − 𝛽2 𝑥̅ which is not
equal. The variance of 𝛼̂1 and 𝛽̂1 are not identical too.
∑(𝑥𝑖 −𝑥̅ )(𝑦𝑖 −𝑦̅)
(b) 𝑌𝑒𝑠 𝛽2 𝑎𝑛𝑑 𝛼2 are identical because 𝛽̂2 = 𝛼̂2 = 2 and the variance
∑(𝜘𝑖 −𝑥̅ )
is also identical.
(c) 𝑀𝑜𝑑𝑒𝑙 − 𝐼𝐼 is better than model-I. Because the variance of model-II is
𝑣(𝛼̂1 ) = 𝑣(𝑦̅) = 0
𝜎2
𝑣(𝛼̂2 ) = ∑(𝑥 −𝑥̅ )2
𝑖
And the variance of model-I is
2 2
𝜎 𝛴𝑥
𝑣(𝛽̂1 ) = 𝑛𝛴(𝑥 −𝑥̅𝑖 )2
𝑖
𝜎2
𝑣(𝛽̂2 ) = 𝛴(𝑥 −𝑥̅ )2
𝑖