Professional Documents
Culture Documents
Financial Econometrics Lecture 4
Financial Econometrics Lecture 4
Financial Econometrics Lecture 4
Lecture 4
Week 5: July 12, 2024
Last week
We studied the desirable (small sample) properties of the OLS
estimators under Gauss-Markov assumptions, i.e. unbiasedness
and efficiency, as well as learnt how to do hypothesis testing
𝐸 𝛽 = 𝐸 𝑋 ′𝑋 −1 𝑋 ′ 𝑦 = 𝐸 𝑋 ′𝑋 −1𝑋′ 𝑋𝛽 + 𝜀
= 𝐸 𝑋′𝑋 −1(𝑋′𝑋)𝛽 + 𝑋 ′𝑋 −1𝑋′𝜀
=𝐸 𝛽+ −1 𝑋 ′ 𝜀 = 𝛽 + −1 𝑋 ′ 𝐸 𝜀 = 𝛽
′
𝑉 𝛽 =𝐸 −1 𝑋′ss ′𝑋 −1]
𝛽 − 𝛽 𝛽 − 𝛽 = 𝐸[
= 𝑋′ 𝑋 −1𝑋′𝐸 ss ′ 𝑋 𝑋 ′𝑋 −1
= 𝑋 ′ 𝑋 −1 𝑋′𝜎2 𝐼𝑁𝑋 𝑋 ′ 𝑋 −1
−1 (𝑋′ 𝑋) 𝑋 ′ 𝑋 −1
= 𝜎2
−1
= 𝜎2
1. Consistency
2. Asymptotic Normality
3. Asymptotic Efficiency
Consistency
Unbiasedness of estimators, though important, cannot always be
achieved
• This means that if we collect more and more data, then we can
get our estimator very close to 𝛽1
From Wooldridge p. 168
Confusion may arise because in reality, we only have a fixed
sample size when we are doing applied work!
The main idea is that obtaining more and more data should get us
very close to the true parameter value of interest
Consistency is based on the Law of Large Numbers (LLN), which
we encountered last week
In the extreme case that the two variables move exactly together (i.e.
one explanatory variable is an exact linear combination of one or
more explanatory variable), it becomes impossible to identify the
individual effects
This is known as perfect collinearity. Technically, this implies that
the (𝑋′𝑋) matrix is not invertible
Hence, t-statistics maybe quite small and you will think the
variable is not significant when in fact it is. F-statistics may be
very large though, as is 𝑅2
All models contain some multicollinearity – the question is when
is it a problem
But we estimate
𝑦 = 𝛽0 + 𝛽1𝑥1 + 𝛽2𝑥2 + s
What if this is not appropriate? i.e. what if the true model is a non-
linear one?
But that need not be fatal for the OLS estimator, because it is still
unbiased
Consequences of heteroscedasticity:
OLS estimator is still unbiased and consistent, but inefficient (no
longer BLUE) and no longer asymptotically efficient
• Unbiased because 𝑋’s are still exogenous (A2 still holds)
• Estimators of the error variance are biased. This invalidates
inference because we can’t trust our OLS estimates for higher
values of 𝑋 since those values are associated with bigger errors
• Inefficiency is harder to demonstrate, but simply put, OLS gives
equal weight to all observations but if some have higher variance
than others, then it would be better to attach low weights to the
ones with higher variance
Since it is only the standard errors rather than the coefficients that
are biased due to heteroscedasticity, one cure can be to find
alternative estimates of the standard errors
White (1980) suggested that replacing 𝜎𝑖2 with s^𝑖2, i.e. the
appropriate OLS residual in the standard formula for calculating
OLS variance-covariance matrix (details not necessary)
For example, if you are working with a dataset of firms, you would
cluster your standard errors by firm (the unit of analysis)
The intuition is that observations from the same firm must have
errors plucked from the same error distribution. Different firms will
then have different error distributions and this is where the
heteroscedasticity comes from
Estimation in presence of heteroscedasticity
Use of Generalized Least Squares (GLS)
The starred model (*) satisfies the classical assumptions about the
error term and we could apply OLS to it obtain estimators that are
BLUE (for more details, see Wooldridge p. 277)
Final point: Robust standard errors approach is used much more in
practice. This is because GLS requires us to know something about
the structure of heteroscedasticity, whereas the robust approach just
estimates the structure from the data
Conclusion
Today, we studied asymptotic properties of OLS estimators