Professional Documents
Culture Documents
4 - 5947019371120429002 Econo ch3
4 - 5947019371120429002 Econo ch3
Lemi Taye
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 1 / 129
Overview
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 2 / 129
Motivation for Multiple Regression
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 3 / 129
The Model with Two Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 4 / 129
The Model with Two Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 5 / 129
The Model with Two Independent Variables
y = β0 + β1 x1 + β2 x2 + u, (2)
where
β0 is the intercept.
β1 measures the change in y with respect to x1 , holding other factors
fixed.
β2 measures the change in y with respect to x2 , holding other factors
fixed.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 6 / 129
The Model with Two Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 7 / 129
The Model with Two Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 8 / 129
The Model with Two Independent Variables
E(u|x1 , x2 ) = 0. (4)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 9 / 129
The Model with Two Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 10 / 129
The Model with k Independent Variables
y = β0 + β1 x1 + β2 x2 + β3 x3 + · · · + βk xk + u, (5)
where
β0 is the intercept.
β1 is the parameter associated with x1 .
β2 is the parameter associated with x2 , and so on.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 11 / 129
The Model with k Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 12 / 129
The Model with k Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 13 / 129
The Model with k Independent Variables
E(u|x1 , x2 , . . . , xk ) = 0. (6)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 14 / 129
The Model with k Independent Variables
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 15 / 129
Obtaining the OLS Estimates
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 16 / 129
Obtaining the OLS Estimates
X
(yi − βb0 − βb1 xi1 − · · · − βbk xik ) = 0
X
xi1 (yi − βb0 − βb1 xi1 − · · · − βbk xik ) = 0
X
xi2 (yi − βb0 − βb1 xi1 − · · · − βbk xik ) = 0 (8)
..
.
X
xik (yi − βb0 − βb1 xi1 − · · · − βbk xik ) = 0.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 17 / 129
Obtaining the OLS Estimates
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 18 / 129
Interpreting the OLS Regression Equation
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 19 / 129
Interpreting the OLS Regression Equation
The estimates βb1 and βb2 have partial effect, or ceteris paribus,
interpretations. From equation (9), we have
∆b
y = βb1 ∆x1 + βb2 ∆x2 ,
so we can obtain the predicted change in y given the changes in x1
and x2 .
In particular, when x2 is held fixed, so that ∆x2 = 0, then
∆b
y = βb1 ∆x1 ,
holding x2 fixed.
The key point is that, by including x2 in our model, we obtain a
coefficient on x1 with a ceteris paribus interpretation. This is why
multiple regression analysis is so useful. Similarly,
∆b
y = βb2 ∆x2 ,
holding x1 fixed.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 20 / 129
Interpreting the OLS Regression Equation
Example (Determinants of College GPA)
Using the data in GPA1 we obtain the following OLS regression line to
predict college GPA from high school GPA and achievement test score:
More interesting estimates are the slope coefficients on hsGP A and ACT .
As expected, there is a positive partial relationship between colGP A and
hsGP A: Holding ACT fixed, another point on hsGP A is associated with
.453 of a point on the college GPA, or almost half a point.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 21 / 129
Interpreting the OLS Regression Equation
The sign on ACT implies that, while holding hsGP A fixed, a change in
the ACT score of 10 points affects colGP A by less than one-tenth of a
point. This is a small effect, and it suggests that, once high school GPA is
accounted for, the ACT score is not a strong predictor of college GPA.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 22 / 129
Interpreting the OLS Regression Equation
The case with more than two independent variables is similar. The
OLS regression line in terms of changes is
∆b
y = βb2 ∆x1 ,
holding x2 , x3 , . . . , xk fixed.
Thus, we have controlled for the variables x2 , x3 , . . . , xk when
estimating the effect of x1 on y. The other coefficients have a similar
interpretation.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 23 / 129
Interpreting the OLS Regression Equation
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 24 / 129
Interpreting the OLS Regression Equation
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 25 / 129
Changing More Than One Independent Variable
Simultaneously
or about 2.6%.
Since exper and tenure each increase by one year, we just add the
coefficients on exper and tenure and multiply by 100 to turn the
effect into a percentage.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 26 / 129
OLS Fitted Values and Residuals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 27 / 129
OLS Fitted Values and Residuals
The OLS fitted values and residuals have some important properties
that are immediate extensions from the single variable case:
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 28 / 129
OLS Fitted Values and Residuals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 29 / 129
Goodness-of-Fit
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 30 / 129
Goodness-of-Fit: Adjusted R-Squared
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 31 / 129
Goodness-of-Fit: Adjusted R-Squared
b2i
P
Intuitively, it is clear that as the number of x variables increases, u
is likely to decrease; hence R2 as defined in (10) will increase.
In view of this, in comparing two regression models with the same
dependent variable but differing number of x variables, one should be
very wary of choosing the model with the highest R2 .
To compare two R2 terms, one must take into account the number of
x variables present in the model.
This can be done readily if we consider an alternative coefficient of
determination, which is as follows:
[SSR/(n − k − 1)]
R̄2 = 1 − (11)
[SST /(n − 1)]
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 32 / 129
Goodness-of-Fit: Adjusted R-Squared
The term adjusted means adjusted for the df associated with the sums
of squares entering into (10).
The primary attractiveness of R̄2 is that it imposes a penalty for
adding additional independent variables to a model.
It is easy to see that R̄2 and R2 are related because, substituting (10)
into (11), we obtain
n−1
R̄2 = 1 − (1 − R2 ) . (12)
n−k−1
It is immediately apparent from (12) that
1 for k > 1, R̄2 < R2 which implies that as the number of x variables
increases, the adjusted R2 increases less than the unadjusted R2 ; and
2 R̄2 can be negative, although R2 is necessarily nonnegative. In case
R̄2 turns out to be negative in an application, its value is taken as
zero.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 33 / 129
The Gauss-Markov Assumptions for Multiple Regression
y = β0 + β1 x1 + β2 x2 + · · · + βk xk + u,
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 34 / 129
The Gauss-Markov Assumptions for Multiple Regression
Assumption MLR.3 (No perfect Collinearity)
In the sample (and therefore in the population), none of the independent
variables is constant, and there are no exact linear relationships among the
independent variables.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 36 / 129
The Gauss-Markov Assumptions for Multiple Regression
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 37 / 129
Unbiasedness of the OLS Estimators
UNBIASEDNESS OF OLS
Under Assumptions MLR.1 through MLR.4,
E(βbj ) = βj , j = 0, 1, . . . , k,
for any values of the population parameter βj . In other words, the OLS
estimators are unbiased estimators of the population parameters.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 38 / 129
The Variance of the OLS Estimators
σ2
Var(βbj ) = , (13)
SSTj (1 − Rj2 )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 39 / 129
The Components of the OLS Variances
Equation (13) shows that the variance of βbj depends on three factors:
σ 2 , SSTj , and Rj2 .
The Error Variance, σ 2
From equation (13), a larger σ 2 means larger sampling variances for
the OLS estimators.
More “noise” in the equation (a larger σ 2 ) makes it more difficult to
estimate the partial effect of any of the independent variables on y,
and this is reflected in higher variances for the OLS slope estimators.
Because σ 2 is a feature of the population, it has nothing to do with
the sample size.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 40 / 129
The Components of the OLS Variances
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 41 / 129
The Components of the OLS Variances
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 42 / 129
Estimating σ 2
df = n − (k + 1)
= (number of observations) − (number of estimated parameters).
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 43 / 129
Standard Errors of the OLS Estimators
b2 , denoted σ
The positive square root of σ b, is called the standard
error of the regression (SER).
For constructing confidence intervals and conducting tests, we will
need to estimate the standard deviation of βbj , which is just the
square root of the variance:
σ
sd(βbj ) = h i1/2 .
SSTj (1 − Rj2 )
σ
se(βbj ) = h
b
i1/2 .
SSTj (1 − Rj2 )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 44 / 129
Efficiency of OLS: The Gauss-Markov Theorem
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 45 / 129
Efficiency of OLS: The Gauss-Markov Theorem
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 46 / 129
Sampling Distributions of the OLS Estimators
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 47 / 129
Sampling Distributions of the OLS Estimators
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 48 / 129
Sampling Distributions of the OLS Estimators
y|x ∼ Normal(β0 + β1 x1 + β2 x2 + β3 x3 + · · · + βk xk , σ 2 ),
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 49 / 129
Sampling Distributions of the OLS Estimators
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 50 / 129
Sampling Distributions of the OLS Estimators
βbj − βj
∼ Normal(0, 1).
sd(βbj )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 52 / 129
Sampling Distributions of the OLS Estimators
The second part of this theorem follows immediately from the fact
that when we standardize a normal random variable by subtracting off
its mean and dividing by its standard deviation, we end up with a
standard normal random variable.
In addition to (15), any linear combination of the β0 , β1 , . . . , βk is
also normally distributed. These facts underlie the testing results in
the remainder of this chapter.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 53 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 54 / 129
Testing Hypotheses: The t Test
βbj − βj
∼ tn−k−1 = tdf , (16)
se(βbj )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 55 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 56 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 57 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 58 / 129
Testing Hypotheses: The t Test
βbj
tβbj ≡ . (18)
se(βbj )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 59 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 60 / 129
Testing Hypotheses: The t Test
A sample value of βbj very far from zero provides evidence against
H0 : βj = 0.
However, we must recognize that there is a sampling error in our
estimate βbj , so the size of βbj must be weighed against its sampling
error.
Since the standard error of βbj is an estimate of the standard deviation
of βbj , t b measures how many estimated standard deviations βbj is
βj
away from zero.
Values of tβbj sufficiently far from zero will result in a rejection of H0 .
The precise rejection rule depends on the alternative hypothesis and
the chosen significance level of the test.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 61 / 129
Testing Hypotheses: The t Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 62 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 63 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 64 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 65 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 66 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 67 / 129
Testing against One-Sided Alternatives
.0041
texper = ≈ 2.41,
.0017
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 68 / 129
Testing against One-Sided Alternatives
H1 : βj < 0, (21)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 70 / 129
Testing against One-Sided Alternatives
We will control for two other factors, average annual teacher compensation
(totcomp) and the number of staff per one thousand students (staf f ).
Teacher compensation is a measure of teacher quality, and staff size is a
rough measure of how much attention students receive.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 72 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 73 / 129
Testing against One-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 74 / 129
Testing against Two-Sided Alternatives
H1 : βj 6= 0. (22)
|tβbj | > c.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 75 / 129
Testing against Two-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 76 / 129
Testing against Two-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 77 / 129
Testing against Two-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 78 / 129
Testing against Two-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 79 / 129
Testing against Two-Sided Alternatives
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 80 / 129
Testing Other Hypotheses about βj
H0 : βj = aj
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 81 / 129
Testing Other Hypotheses about βj
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 82 / 129
Computing p-Values for t Tests
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 83 / 129
Computing p-Values for t Tests
That is, the p-value is the significance level of the test when we use
the value of the test statistic, 1.85 in the above example, as the
critical value for the test.
This p-value is shown in Figure 5.
Because a p-value is a probability, its value is always between zero and
one.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 84 / 129
Computing p-Values for t Tests
In the example with df = 40 and t = 1.85, the p-value for testing the
null hypothesis H0 : βj = 0 against the two-sided alternative is
computed as
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 86 / 129
Computing p-Values for t Tests
The previous example illustrates that once the p-value has been
computed, a classical test can be carried out at any desired level.
If α denotes the significance level of the test (in decimal form), then
H0 is rejected if p-value < α; otherwise, H0 is not rejected at the
100 ·α% level.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 87 / 129
Computing p-Values for t Tests
We can use the scrap rate to measure the effect of worker training on
productivity. Using the data in JTRAIN, but only for the year 1987 and for
nonunionized firms, we obtain the following estimated equation:
The variable hrsemp is annual hours of training per employee, sales is annual
firm sales (in dollars), and employ is the number of firm employees.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 88 / 129
Computing p-Values for t Tests
The p-value is easily computed as P (T25 < −1.26) = .110. Thus, hrsemp
is not statistically significant, even at the 10% level.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 89 / 129
Confidence Intervals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 90 / 129
Confidence Intervals
More precisely, the lower and upper bounds of the confidence interval
are given by
β j ≡ βbj − c · se(βbj )
and
β¯j ≡ βbj + c · se(βbj ),
respectively.
What is the meaning of a confidence interval? If random samples were
obtained over and over again, with β j and β¯j computed each time,
then the (unknown) population value βj would lie in the interval
(β j , β¯j ) for 95% of the samples.
Three quantities are needed to construct a confidence interval: βbj ,
se(βbj ), and c.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 91 / 129
Confidence Intervals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 92 / 129
Confidence Intervals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 93 / 129
Confidence Intervals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 94 / 129
Confidence Intervals
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 95 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
In applications, we must often test hypotheses involving more than
one of the population parameters.
In this section, we show how to test a single hypothesis involving more
than one of the βj .
To illustrate the general approach, we will consider a simple model to
compare the returns to education at junior colleges and four-year
colleges; for simplicity, we refer to the latter as “universities.”
The population includes working people with a high school degree,
and the model is
log(wage) = β0 + β1 jc + β2 univ + β3 exper + u, (24)
where
jc = number of years attending a two-year college.
univ = number of years at a four-year college.
exper = months in the workforce.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 96 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
H0 : β1 = β2 (25)
H1 : β1 < β2 (26)
We cannot simply use the individual t statistics for βb1 and βb2 to test
H0 .
However, conceptually, there is no difficulty in constructing a t
statistic for testing (25).
To do so, we rewrite the null and alternative as H0 : β1 − β2 = 0 and
H1 : β1 − β2 < 0, respectively.
The t statistic is based on whether the estimated difference βb1 − βb2 is
sufficiently less than zero to warrant rejecting (25) in favor of (26).
To account for the sampling error in our estimators, we standardize
this difference by dividing by the standard error:
βb1 − βb2
t= . (27)
se(βb1 − βb2 )
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 98 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 99 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
Example
Using the data in TWOYEAR, which comes from Kane and Rouse (1995), we
estimate equation (24):
It is clear from this result that jc and univ have both economically and
statistically significant effects on wage. This is certainly of interest, but we are
more concerned about testing whether the estimated difference in the coefficients
is statistically significant. The difference is estimated as βb1 − βb2 = −.0102, is the
numerator of the t statistic in (24). Unfortunately, the regression result do not
contain enough information to obtain the standard error of βb1 − βb2 .
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 101 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
Example (continued )
The standard error in the denominator of the t-statistic is the square root of
\ \ \ \
Var(βb1 − βb2 ) = Var(βb1 ) + Var(βb2 ) − 2Cov(βb1 , βb2 )
If you are using Stata you can compute this value manually using the estimated
covariance matrix of the least squares estimates, obtained post-estimation using
estat vce.
. estat vce
Covariance matrix of coefficients of regress model
e(V) jc univ exper _cons
jc .00004663
univ 1.928e-06 5.330e-06
exper -1.718e-08 3.933e-08 2.480e-08
_cons -.00001741 -.00001573 -3.105e-06 .00044353
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 102 / 129
Testing Hypotheses about a Single Linear Combination of
the Parameters
Example (continued )
Using lincom we can estimate a linear combination such as c1 β1 + c2 β2 , and test
the more general form of linear hypotheses quite easily. For this particular
example, c1 = 1 and c2 = −1. In the case of the command lincom we can
simply refer to the variable names, such as jc - univ. The command will
compute the value of the expression and its standard error, and produce a t
statistic and an interval estimate.
. lincom jc-univ
( 1) jc - univ = 0
The t statistic for testing (25) is −1.47. Against the one-sided alternative (26),
the p-value is about .071, so there is some, but not strong, evidence against (25).
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 103 / 129
Testing Multiple Linear Restrictions: The F Test
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 104 / 129
The F Test: Testing Exclusion Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 105 / 129
The F Test: Testing Exclusion Restrictions
H0 : β2 = 0, β3 = 0
H1 : β2 6= 0 or β3 6= 0 or both are nonzero
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 106 / 129
The F Test: Testing Exclusion Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 107 / 129
The F Test: Testing Exclusion Restrictions
Using the data in the file ANDY to estimate (28) and (29), we find
that SSRu = 1532.084 and SSRu = 1896.391.
Adding advert and advert2 to the equation reduces the sum of
squared errors from 1896.391 to 1532.084.
What the F test does is to assess whether this reduction is sufficiently
large to be significant.
If adding the extra variables has little effect on the sum of squared
errors, then those variables contribute little to explaining variation in
the dependent variable, and there is support for a null hypothesis that
drops them.
On the other hand, if adding the variables leads to a big reduction in
the sum of squared errors, those variables contribute significantly to
explaining the variation in the dependent variable, and we have
evidence against the null hypothesis.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 108 / 129
The F Test: Testing Exclusion Restrictions
H0 : βk−j+1 = 0, . . . , βk = 0, (31)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 109 / 129
The F Test: Testing Exclusion Restrictions
y = β0 + β1 x1 + · · · + βk−j xk−j + u.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 110 / 129
The F Test: Testing Exclusion Restrictions
F ∼ F(j, n−k−1)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 111 / 129
The F Test: Testing Exclusion Restrictions
If the null hypothesis is not true, then the difference between SSRr
and SSRu becomes large, implying that the restrictions placed on the
model by the null hypothesis significantly reduce the ability of the
model to fit the data.
A large value for SSRr − SSRu means that the value of F tends to
be large, so that we reject the null hypothesis if the value of the F
test statistic becomes too large.
What is too large is decided by comparing the value of F to a critical
value c, which leaves a probability α in the upper tail of the F
distribution with j and n − k − 1 degrees of freedom.
The 10%, 5%, and 1% critical values for the F distribution are
tabulated.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 112 / 129
The F Test: Testing Exclusion Restrictions
F > c.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 113 / 129
The F Test: Testing Exclusion Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 114 / 129
The F Test: Testing Exclusion Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 115 / 129
The F Test: Testing Exclusion Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 116 / 129
The F Test: Testing Exclusion Restrictions
Since F = 8.44 is well above the 1% critical value, we reject the null
hypothesis that both β2 = 0 and β3 = 0, and conclude that at least
one of them is not zero. Advertising does have a significant effect
upon sales revenue.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 117 / 129
The R-Squared Form of the F Statistic
p-value = P (F > F ),
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 119 / 129
The F Statistic for Overall Significance of a Regression
This null hypothesis states that none of the explanatory variables has
an effect on y.
Stated in terms of the parameters, the null is that all slope parameters
are zero:
H0 : β1 = β2 = · · · = βk = 0, (34)
and the alternative is that at least one of the βj is different from zero.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 120 / 129
The F Statistic for Overall Significance of a Regression
There are k restrictions in (34), and when we impose them, we get the
restricted model
y = β0 + u; (35)
all independent variables have been dropped from the equation.
Now, the R-squared from estimating (35) is zero; none of the variation
in y is being explained because there are no explanatory variables.
Therefore, the F statistic for testing (34) can be written as
R2 /k
, (36)
(1 − R2 )/(n − k − 1)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 121 / 129
The F Statistic for Overall Significance of a Regression
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 122 / 129
Testing General Linear Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 123 / 129
Testing General Linear Restrictions
where
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 124 / 129
Testing General Linear Restrictions
Now, suppose we would like to test whether the assessed housing price
is a rational valuation.
If this is the case, then a 1% change in assess should be associated
with a 1% change in price; that is, β1 = 1.
In addition, lotsize, sqrf t, and bdrms should not help to explain
log(price), once the assessed value has been controlled for.
Together, these hypotheses can be stated as
H0 : β1 = 1, β2 = 0, β3 = 0, β4 = 0. (38)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 125 / 129
Testing General Linear Restrictions
y = β0 + β1 x1 + β2 x2 + β3 x3 + β4 x4 + u, (39)
y − x1 = β0 + u. (40)
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 126 / 129
Testing General Linear Restrictions
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 127 / 129
Testing General Linear Restrictions
The SSR from the restricted model turns out to be SSRr = 1.880,
and so the F statistic is [(1.880 − 1.822)/1.822](83/4) = .661.
The 5% critical value in an F distribution with (4, 83) df is about
2.50, and so we fail to reject H0 .
There is essentially no evidence against the hypothesis that the
assessed values are rational.
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 128 / 129
************* End of Chapter Three *************
Lemi Taye (AAUSC) Ch 3: Multiple Linear Regression December 14, 2019 129 / 129