Professional Documents
Culture Documents
Econometrics Summary
Econometrics Summary
Model: y= ß0 + ß1x + u
Ordinary Least Squares (OLS): estimate ß0,ß1 by minimizing sum of the squared residuals:
- Fitted value for y when x = xi: y^i = ß^0 + ß^1xi
- Residual= Di erence between actual yi and its tted value: ûi = yi - y^i = yi- ß^0-ß^1xi
- Min ∑ ûi2 = ∑(yi- ß^0 - ß^1xi)2
• In SLR A4 is often unrealistic => control for more factors (take them „out of the error term“ and
control for them, when thinking its correlated with regressor)
• ßi measures the ceteris paribus e ect of x on y, when holding the other regressors constant
• Assume SLR: y~= ß~0 + ß~1x and MLR: y^ = ß^0 + ß^1x1 + ß^2x2 => in general: ß~1 ≠ ß^1
unless: ß^2 = 0 (no partial e ect of x2) or x1 and x2 are uncorrelated in the sample
• Assumptions: (A1) Linear in parameters (linear in ßs, xi can enter non-linearly), (A2) Random
sampling, (A3) no perfect collinearity (no exact linear relationship among independent variables),
(A4) zero conditional mean (E(uIx1,x2,x3,…,xk)= 0)
Hypothesis testing:
t-Test:
• T-distributed with H0: ßj=0 and H1: ßj ≠ 0 (meaning x has no e ect on y, controlling for other x’s)
F-test:
• F-test is a joined hypothesis test that simultaneously tests two or more restrictions under H0
(can also be used for testing one restriction with more than one parameter: i.e. H0: ß1=ß3)
• Test: H0: Rß-r = 0 vs. H1: Rß-r ≠ 0 => Weighted squared distance measure for H0
• Heteroskedasticity robust F- statistic: R = q x (k+1) Matrix with q≤k+1 and rg(R) = q, r is a q x 1 vector
Sum of squared distances for q restrictions of H0
Estimator for (approximate) variance matrix for (Rß^-r)
- Estimator for variance matrix for (Rß^-r) as the weighted sum of the squared distances
(=> restrictions with high variance have a relatively lower weighting)
- Considering covariances between restrictions (if we have covariances=> no diagonal matrix):
With ai measuring the distances of the restrictions:
Rß^ -r = a ≠ 0
- Inverse: we divide by diagonal entries, i.e. the distances with a high estimated variance
get a lower weight, relative to the distances that are estimated more accurately, i.e. with
smaller variance
fi
fi
fi
fi
fi
ff
Testing linear Combinations:
• Instead of testing if ß1 is equal to a constant, test if its equal to another parameter: H0: ß1=ß2
• t- statistic:
• Jointly test multiple hypotheses about parameters (e.g. H0: ßk-q+1=0,…,ßk=0, H1: H0 not true)
=> can’t check each separately, because we want to know if they are jointly signi cant at given
level (it’s possible for no parameter to be individually signi cant, while being jointly signi cant)
• Estimate the „restricted model“ without xk-q+1,…,xk as well as the „unrestricted model“ with all
x’s included => Does the change in SSR is big enough to warrant inclusion of xk-q+1,…,xk:
r = restricted model
ur = unrestricted model
q= number of restrictions or dfr-dfur, where dfur = n-k-1
• F statistic is always positive (SSR from the restricted model > SSR from the unrestricted)
• F statistic measures relative increase in SSR when moving from unrestricted to restricted model
• F~Fq,n-k-1, where q is referred to as numerator of freedom and n-k-1 as dominator degrees of
freedom => Reject H0 at a signi cance level α if F > c
• Alternative use:
Calculate p-values by looking up the
percentile in the appropriate F distribution
• If only one exclusion is being tested: F = t2, and the p-values will be the same
Asymptotics:
• If we have consistency and n is large, then t and F statistics will be approximately t- and F-
distributed (under Gauss Markov assumptions, the OLS estimator is consistent and unbiased)
• Standardized average of any population with mean µ and variance σ2 is asymptotically ~ N(0,1):
OLS estimators are asymptotically normal
Elasticity:
• Idea: replace y and each x variable with a standardized version i.e. subtract mean and divide by
standard deviation (=> variables are distributed with mean 0 and s.d.1)
=> coe cient re ects change in standard deviation of y for a one standard deviation change in x
• Changing the scale of the y variable will lead to a corresponding change in the scale of the
coe cients and standard errors, no change in the signi cance or interpretation
• Changing the scale of an x variable will lead to a change in the scale of that coe cient and
standard error, no change in signi cance or interpretation.
• Functional forms: OLS can also be used for not strictly linear relationships in x and y by using
nonlinear functions (log, quadratic, interactions) of x and y (model is still linear in its parameters)
- Quadratic models: For a model y=ß0+ß1x+ß2x2+u => change in y: dy/dx = ß1 + 2ß2x
- Interaction terms: y=ß0+ß1x1+ß2x2+ß3x1x2+u => change in y: dy/dx = ß1 + ß3x2
Log Models:
Chow-Test:
• Goal: Testing whether a regression function is di erent for one group versus another can be
thought of as simply testing for the joint signi cance of the dummy and its interactions with all
other x variables (e.g. di erence for man and woman).
- Test fro structural change over time: Interact X variables whose e ect we suspect to change
over time with time indicators (e.g. year dummies) or test for individual (t-test) or joint (F-test)
signi cance of coe cients
• If we run the restricted model for group one and get SSR1, then for group two and get SSR2
• Run the restricted model for all to get SSR:
SSR from model: y = Xß
SSR1 from model Mmale: y = Xßmale
SSR2 from model: Mwoman: y= Xßwoman
• The Chow test is just a F test for exclusion restrictions => SSRur= SSR1+ SSR2
• Note: Restrictions: k+1 (each of slope coe cients and intercept)
• Note: unrestricted model would estimate 2 di erent intercepts and 2 di erent slope coe cients:
df=n-2k-2
Heteroskedasticity:
• Def: variance of u is di erent for di erent values of the x’s, then the errors are heteroskedastic
• OLS coe cients are still unbiased and consistent, but the estimators of coe cient variances
Var(ß^j) are biased if we have heteroskedasticity
• Estimator for Var(ß^1) for SLR:
• Test H0: Var(uIx1,x2,…,xk) = σ2, which is equivalent to H0: E(u2 I x1,x2,…,xk) = E(u2) = s2
• If we assume the relationship between u2 and xj to be linear, we can test it as a linear restriction:
u2 = d0 + d1x1 + … + dkxk + v => H0: d1 = d2 = … = dk = 0
White Test:
• Goal: test model for misspeci cation => test for hetersoskedasticity
• Idea: White test allows for nonlinearities by using squares and cross products of all the x’s
- test: H0: Var(ui I Xi) = σ2 vs H1: Var(ui I Xi) = σ2i (variance depends on regressors)
- Use û2 as proxy for σ2 and then test if conditional variance depends linearly on the
regressors and the squares of the regressors (standard case)
- If there is a relationship between squared residuals and the speci c function of the
regressors we have evidence against homoscedasticity (reject H0)
• Procedure: Test if σ2i = f(X1i,…,Xki)
1. Run the original OLS model, save the residuals, ûi: Yi-Xi’ß^
2. Determine R2 of the auxiliary regression/test regression:
• In case of heteroskedasticity robust standard errors for OLS estimates still unbiased&consistent,
but WLS/GLS is more e cient if we know smth about speci c form of heteroskedasticity
• Idea: transform model into one that has homoskedastic errors - called weighted least squares
• Suppose heteroskedasticity is modeled as Var(uIx) = σ2h(x) => gure out what h(x) = hi looks like
- E(ui/√hi Ix)= 0, because hi is only a function of x and Var(ui/√hi Ix)= σ2, because we know σ2hi
- If we divided our whole equation by √hi => model where the error is homoskedastic
ffi
fi
fi
ffi
ffi
fi
fi
fi
fi
fi
fi
Generalized Least Squares:
• Estimating the transformed equation by OLS => example of generalized least squares (GLS):
• GLS is a weighted least squares (WLS) procedure where each squared residual is weighted by
the inverse of Var(ui I xi) (OLS as special case of GLS, where all weights are equal)
• If data is heteroskedastic => GLS is more e cient than OLS (considers heteroskedasticity)
• Typically we don’t know the form of heteroskedasticity => estimate h(xi):
- Model: Var(uIx) = σ2 ed0+d1x1+…+dkxk (if we know di we could apply WLS => estimate di)
(used exponential form, because estimated variances must be positive in order to perform WLS)
1. Run the original OLS model, save the residuals, û, square them:
u2 = Var(uIx) = σ2 ed0+d1x1+…+dkxk v , where E(vIx) = 1, if E(v) = 1 (we can write e)
2. Take the log of the squared residuals:
ln(u2) = a0 + d1x1 + … + dkxk + e , where E(e) = 0 and e is independent of x
=> û is an estimate of u, so we can estimate this by OLS
3. Regress them on all of the independent variables and get the tted values g^
4. Do WLS using 1/exp(g^) as the weight
=> Estimate of h is obtained as h^=exp(g^), where g^ are the residuals from the regression
of log(û2) on the xk => Inverse of this is the weight
=> WLS and GLS are equal if we don’t have autocorrelation (diagonal psi matrix)
• Goal: test of functional form of the model to nd the right functional form for our model
- Are there (non)linear relationships, should we use log, quadratic forms or interaction terms?
• Idea: Instead of adding functions of the x’s directly, we add and test functions of y^
=> estimate: y= ß0 + ß1x1 + … + ßkxk + d1y2^ + d2y3^ + error (with Yi^=ß0^+ß1^x1+ßk^xk)
=> test: H0: d1= 0, d2= 0 (linear model is correctly speci ed) vs: H1: minimum of one dj ≠ 0
=> test statistic: F~F2,n-k-3
=> if p-value > signi cance level: Accept H0 => linearity of regressors
=> if p-value < signi cance level: Reject H0 => non linearity of regressors /omitted variables
• Rejection of H0 is evidence for functional form misspeci cation, but doesn’t tell correct form
TO DO Wilcox test:
fi
fi
fi
ffi
fi
fi
fi
fi