Professional Documents
Culture Documents
Econometrics Course for RDAE Chapter 5
Econometrics Course for RDAE Chapter 5
College of Agriculture
Department of Rural development
&
Agricultural Extension
Course Name: Econometrics
March, 2024
Mersa
Chapter 5
Econometric Problems
Outline of the chapter
5.1. Heteroskedastcity
5.2. Autocorrelation
5.3. Multicollinearity
5.4. Non-normality
Assumptions Revisited
X i.e. E(ui) = 0
OLS residuals, ei, since these are the once we observe and
ln i ln 2 ln X i vi
2
ln ei ln 2 ln X i vi
2
since ² is not known.
Recall that:
di 2
rS 1 6
N ( N 1)
2
variable separately.
e) Goldfeld and Quandt Test
This is the most popular test and usually suitable for large
samples.
If it is assumed that the variance (i²) is positively related to
one of the explanatory variables in the regression model and,
If the number of observations is at least twice as many as the
parameters to be estimated, the test can be used.
Given the model
Yi 0 1 X i U i
i Xi
2 2 2
Goldfeld and Quandt suggest the following steps:
Rss 2 / df ( n c 2k )
F ~ FV1V2 v1 v2
Rss1 / df 2
If the two variances tend to be the same, then F approaches
unity. If the variances differ we will have values for F different
from one.
The higher the F-ratio, the stronger the evidence of
heteroscedasticity.
Note: There are also other methods of testing the existence of
heteroscedasticity in your data.
These are Glejser Test, Breusch-Pagan-Godfrey Test, White’s General Test
and Koenker-Bassett Test, the details for which you are supposed to refer.
Remedial Measures
If var(U i ) i X i , then
2 2
E (U i ) X 1i
2 2 2
1 V=
0 ( ) 1 Vi
X 1i
Now,
E(= E()2 = E() =
Yi 0 X 1i Ui
1
X 1i X 1i X 1i X 1i
1
0 1 X 1i Vi
X 1i
E (U i ) E (Yi )
2 2 2
Ui
Vi
E (Yi )
1 X 1i
0 1 Vi
E (Yi ) E (Yi )
E (Y )
i E (Y i ) 2 i
E (Y i ) 2 i
Serial correlation implies that the error term from one time
period depends in some systematic way on error terms from
other time periods.
Autocorrelation is more a problem of time series data than
cross_sectional data.
If by chance, such a correlation is observed in cross-sectional
units, it is called spatial autocorrelation.
,ij
E (U iU j ) 0 ,ij
E (U t ) 0
E (U t ,U t 1 ) 0 t t 1
U t ~ N (0, 2 )
Consequences of serial correlation
When the disturbance term exhibits serial correlation, the values
as well as the standard errors of the parameters are affected.
1) The estimates of the parameters remain unbiased even in
the presence of autocorrelation but the X’s and the u’s must be
uncorrelated.
2) Serial correlation increases the variance of the OLS estimators.
The minimum variance property of the OLS parameter estimates
is violated.
That means the OLS are no longer efficient.
w ith o u t s e ria l
c o rre la tio n
~
Var ( ˆ ) Var ( )
w ith s e ria l
c o rre la tio n
ˆ
Figure 4.1: The distribution of with and without serial correlation.
3. Due to serial correlation the variance of the disturbance term,
Ui may be underestimated.
This problem is particularly pronounced when there is positive
autocorrelation.
4. If the Uis are autocorrelated, then prediction based on the OLS
estimates will be inefficient. This is because of larger variance of
the parameters.
Since the variances of the OLS estimators are not minimal as
compared with other estimators, the standard error of the
forecast from the OLS will not have the least value.
Detecting Autocorrelation
Some rough idea about the existence of autocorrelation may be
gained by plotting the residuals either against their own
lagged values or against time.
et
et
et = +ve et = +ve
et-1 = -ve et-1 = +ve
et-1
et-1 et = -ve
et-1 = -ve
et-1 = +ve
et = -ve
U t PU t 1 Et then Et PEt 1 U t
The test may be outlined as
HO : P 0
H1 : P 0
This test is, however, applicable where the underlying
assumptions are met:
The regression model includes an intercept term
The serial correlation is first order in nature
The regression does not include the lagged dependent
variable as an explanatory variable
There are no missing observations in the data
The equation for the Durban-Watson d statistic is:
N
t t 1
( e e ) 2
d t 2
N
et
2
t 1
Which is simply the ratio of the sum of squared differences in
successive residuals to the RSS
Note that the numerator has one fewer observation than the
denominator, because an observation must be used to calculate
A great advantage of the d-statistic is that it is based on the
estimated residuals.
Thus it is often reported together with R², t, etc.
The d-statistic equals zero if there is extreme positive serial
.
correlation, two if there is no serial correlation, and four if
there is extreme negative correlation.
1. Extreme positive serial correlation: d 0
et et 1 so (et et 1 ) 0 and d 0
d
t
( 2 e ) 2
and d 4
et
2
3. No serial correlation: d 2
t t 1 et et 1 2 et et 1
2 2
( e e ) 2
d 2
e e
2 2
t t
e e
2 2
Since t and t 1 differ in only one observation,
t t 1 et et 1 2 et et 1
2 2
( e e ) 2
Thus, d
e e
2 2
t t
2(1
eet t 1
)
e e e
2
t 1 t t 1 ˆ
P
d 2(1 Pˆ ) since
e t 1
2
But, since –1 P 1 the above identity can be written as: 0 d
4
Therefore, the bounds of d must lie within these limits.
Reject H0 Reject H0
+ve -ve
autocorr. autocorr.
accept H0
no serial
correlation
0 d
dL dU 4-dU 4-dL 4
If Pˆ 0 d = 2, no serial autocorrelation.
If Pˆ 1 d = 0, evidence of positive autocorrelation.
If Pˆ 1 d = 4, evidence of negative autocorrelation
w
iths eve re
m
u ltic
ollinearity
̂
(3) The computed t-ratios will fall i.e. insignificant t-ratios will
be observed in the presence of multicollinearity.
t since SE ( ˆ ) increases t-falls.
SE ( ˆ )
Thus, one may increasingly accept the null hypothesis that the
relevant true population’s value is zero
H0 : i 0
X t X t X t 1
If an equation (or some of the variables in an equation) is
switched from its normal specification to a first difference
specification, it is quite likely that the degree of
multicollinearity will be significantly reduced for two
reasons.
Since multicollinearity is a sample phenomenon, any change
in the definitions of the variables in that sample will change
the degree of multicollinearity.
Multicollinearity takes place most frequently in time-series
data, in which first differences are far less likely to move
steadily upward than are the aggregates from which they
are calculated.
(4) Increase the sample size
Another solution to reduce the degree of multicollinearity is
to attempt to increase the size of the sample.
Larger data set (often requiring new data collection) will allow
more accurate estimates than a small one,
Since the large sample normally will reduce somewhat the
variance of the estimated coefficients reducing the impact
of multicollinearity.
But, for most economic and business applications, this solution
is not feasible.
As a result new data are generally impossible or quite
expensive to find.
One way to increase the sample is to pool cross-sectional and
time series data.
5.4 Non-Normality
One of the key assumptions of OLS is that the residuals of the
model are normally distributed.
The stochastic (disturbance) term ui are normally distributed i.e.,
Ui ~ N(0, ²)
• It is also important to make sure your variables are not skewed too
far either negatively or positively.
• Let’s cover how to examine whether residuals are normally
distributed, as well as how to handle greatly-skewed variables.
• A non-normal distribution is any statistical distribution that is not
normal.
• Non-normal distributions can take many different shapes,
including skewed, bimodal, and uniform.
• Non-normal distributions are more challenging to analyze than
normal distributions because they do not have the same
predictable patterns.
Reasons, why the data are not normally distributed?
1. Extreme values or outliers
1. Two or more process overlapping
2. Insufficient data
3. Subset of the main sample
4. Data follows some other distribution
Consequences of Non-normality