CH 02

You might also like

Download as pdf or txt
Download as pdf or txt
You are on page 1of 137

Chapter 2

The Simple Linear Regression Model:


Specification and Estimation

Walter R. Paczkowski
Rutgers University
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 1
Chapter Contents

 2.1 An Economic Model


 2.2 An Econometric Model
 2.3 Estimating the Regression Parameters
 2.4 Assessing the Least Squares Estimators
 2.5 The Gauss-Markov Theorem
 2.6 The Probability Distributions of the Least
Squares Estimators
 2.7 Estimating the Variance of the Error Term
 2.8 Estimating Nonlinear Relationships
 2.9 Regression with Indicator Variables
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 2
2.1
An Economic Model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 3
2.1
An Economic
Model

As economists we are usually more interested in


studying relationships between variables
– Economic theory tells us that expenditure on
economic goods depends on income
– Consequently we call y the ‘‘dependent
variable’’ and x the independent’’ or
‘‘explanatory’’ variable
– In econometrics, we recognize that real-world
expenditures are random variables, and we
want to use data to learn about the relationship

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 4
2.1
An Economic
Model

The pdf is a conditional probability density


function since it is ‘‘conditional’’ upon an x
– The conditional mean, or expected value, of y is
E(y|x)
• The expected value of a random variable is
called its ‘‘mean’’ value, which is really a
contraction of population mean, the center of
the probability distribution of the random
variable
• This is not the same as the sample mean,
which is the arithmetic average of numerical
values

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 5
2.1
An Economic Figure 2.1a Probability distribution of food expenditure y given
Model income x = $1000

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 6
2.1
An Economic
Model

The conditional variance of y is σ2 which


measures the dispersion of y about its mean μy|x
– The parameters μy|x and σ2, if they were known,
would give us some valuable information about
the population we are considering

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 7
2.1
An Economic Figure 2.1b Probability distributions of food expenditures y
Model given incomes x = $1000 and x = $2000

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 8
2.1
An Economic
Model

In order to investigate the relationship between


expenditure and income we must build an
economic model and then a corresponding
econometric model that forms the basis for a
quantitative or empirical economic analysis
– This econometric model is also called a
regression model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 9
2.1
An Economic
Model

The simple regression function is written as

Eq. 2.1 E(Ey( |yx| )x)μ yy β2 1β2x2 x Eq. 2.1

where β1 is the intercept and β2 is the slope

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 10
2.1
An Economic
Model

It is called E   y  1not
( y | x)regression
simple  because
2x it is easy,
but because there is only one explanatory variable
on the right-hand side of the equation

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 11
2.1
An Economic Figure 2.2 The economic model: a linear relationship between
Model average per person food expenditure and income

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 12
2.1
An Economic
Model

The slope of the regression line can be written as:


E ( y | x) dE ( y | x)
Eq. 2.2 β2  
x dx
E ( y | x) dE ( y | x)
where “Δ” denotes
2  “change
x
 in” and “dE(y|x)/dx”
dx
Eq. 2.2

denotes the derivative of the expected value of y


given an x value
“Δ” denotes “change in” and “dE(y|x)/dx” denotes the derivative of
the expected value of y given an x value

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 13
2.2
An Econometric Model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 14
2.2
An Econometric
Model
Figure 2.3 The probability density function for y at two levels of income

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 15
2.2
An Econometric
Model

There are several key assumptions underlying the


simple linear regression
– More will be added later

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 16
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - I

Assumption 1:
The mean value of y, for each value of x, is
given by the linear regression

E ( y | x)  β1  β2 x

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 17
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - I

Assumption 2:
For each value of x, the values of y are
distributed about their mean value, following
probability distributions that all have the same
variance
var( y | x)  σ 2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 18
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - I

Assumption 3:
The sample values of y are all uncorrelated,
and have zero covariance, implying that there
is no linear association among them
cov( yi , y j )  0
This assumption can be made stronger by
assuming that the values of y are all
statistically independent

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 19
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - I

Assumption 4:
The variable x is not random, and must take at
least two different values

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 20
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - I

Assumption 5:
(optional) The values of y are normally
distributed about their mean for each value of x

y ~ N (β1  β2 x,σ ) 2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 21
2.2
An Econometric
Model

2.2.1
Introducing the
Error Term
The random error term is defined as

Eq. 2.3 e  y  E  y | x   y  β1  β2 x

– Rearranging gives

Eq. 2.4 y  β1  β 2 x  e

where y is the dependent variable and x is the


independent variable

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 22
2.2
An Econometric
Model

2.2.1
Introducing the
Error Term

The expected value of the error term, given x, is

E (e | x)  E ( y | x)  β1  β2 x  0

The mean value of the error term, given x, is zero

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 23
2.2
An Econometric
Model
Figure 2.4 Probability density functions for e and y

2.2.1
Introducing the
Error Term

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 24
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR1:
The value of y, for each value of x, is:

y  β1  β2 x  e

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 25
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR2:
The expected value of the random error e is:

E (e)  0

This is equivalent to assuming that

E ( y)  β1  β2 x

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 26
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR3:
The variance of the random error e is:

var(e)  σ2  var( y)

The random variables y and e have the


same variance because they differ only by
a constant.

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 27
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR4:
The covariance between any pair of random
errors, ei and ej is:
cov(ei , e j )  cov( yi , y j )  0

The stronger version of this assumption is that


the random errors e are statistically independent,
in which case the values of the dependent
variable y are also statistically independent
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 28
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR5:
The variable x is not random, and must take at
least two different values

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 29
2.2
An Econometric
Model
ASSUMPTIONS OF THE SIMPLE LINEAR REGRESSION MODEL - II

2.2.1
Introducing the
Error Term

Assumption SR6:
(optional) The values of e are normally
distributed about their mean if the values of y
are normally distributed, and vice versa

e ~N (0, σ2 )

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 30
2.2
An Econometric
Model
Figure 2.5 The relationship among y, e and the true regression line

2.2.1
Introducing the
Error Term

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 31
2.3
Estimating the Regression Parameters

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 32
2.3
Estimating the
Regression Table 2.1 Food Expenditure and Income Data
Parameters

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 33
2.3
Estimating the
Regression Figure 2.6 Data for food expenditure example
Parameters

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 34
2.3
Estimating the
Regression
Parameters

2.3.1
The Least Squares
Principle

The fitted regression line is:


Eq. 2.5 yˆi  b1  b2 xi
The least squares residual is:

Eq. 2.6 eˆi  yi  yˆi  yi  b1  b2 xi

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 35
2.3
Estimating the
Regression Figure 2.7 The relationship among y, ê and the fitted regression line
Parameters

2.3.1
The Least Squares
Principle

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 36
2.3
Estimating the
Regression
Parameters

2.3.1
The Least Squares
Principle

Suppose we have another fitted line:


yˆi*  b1*  b2* xi

The least squares line has the smaller sum of


squared residuals:
N N
if SSE   eˆi2 and SSE *   eˆi*2 then SSE  SSE *
i 1 i 1

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 37
2.3
Estimating the
Regression
Parameters

2.3.1
The Least Squares
Principle

Least squares estimates for the unknown


parameters β1 and β2 are obtained my minimizing
the sum of squares function:
N
S (β1 ,β 2 )   ( yi  β1  β 2 xi ) 2
i 1

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 38
2.3
Estimating the
Regression THE LEAST SQUARES ESTIMATORS
Parameters

2.3.1
The Least Squares
Principle

b2 
 ( x  x )( y  y )
i i

 (x  x)
Eq. 2.7
2
i

Eq. 2.8 b1  y  b2 x

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 39
2.3
Estimating the
Regression
Parameters

2.3.2
Estimates for the
Food Expenditure


Function
( x  x )( y  y ) 18671.2684
b2  i
i
 10.2096
 (x  x) i
2
1828.7876

b1  y  b2 x  283.5735  (10.2096)(19.6048)  83.4160

A convenient way to report the values for b1


and b2 is to write out the estimated or fitted
regression line:

yˆ i  83.42  10.21xi

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 40
2.3
Estimating the
Regression Figure 2.8 The fitted regression line
Parameters

2.3.2
Estimates for the
Food Expenditure
Function

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 41
2.3
Estimating the
Regression
Parameters

The value b2 = 10.21 is an estimate of 2, the


2.3.3
Interpreting the
Estimates

amount by which weekly expenditure on food per


household increases when household weekly
income increases by $100. Thus, we estimate that
if income goes up by $100, expected weekly
expenditure on food will increase by
approximately $10.21
– Strictly speaking, the intercept estimate b1 =
83.42 is an estimate of the weekly food
expenditure on food for a household with zero
income

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 42
2.3
Estimating the
Regression
Parameters

2.3.3a
Elasticities Income elasticity is a useful way to characterize
the responsiveness of consumer expenditure to
changes in income. The elasticity of a variable y
with respect to another variable x is:

percentage change in y y x
 
percentage change in x x y

In the linear economic model given by Eq. 2.1 we


have shown that
E ( y )
β2 
x
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 43
2.3
Estimating the
Regression
Parameters

2.3.3a
Elasticities The elasticity of mean expenditure with respect to
income is:

E ( y ) E ( y ) E ( y ) x x
Eq. 2.9    β2
x x x E ( y) E ( y)

A frequently used alternative is to calculate the


elasticity at the “point of the means” because it is
a representative point on the regression line.
x 19.60
ˆ  b2  10.21  0.71
y 283.57

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 44
2.3
Estimating the
Regression
Parameters

2.3.3b
Prediction

Suppose that we wanted to predict weekly food


expenditure for a household with a weekly income
of $2000. This prediction is carried out by
substituting x = 20 into our estimated equation to
obtain:
yˆ  83.42  10.21xi  83.42  10.21(20)  287.61

We predict that a household with a weekly income


of $2000 will spend $287.61 per week on food

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 45
2.3
Estimating the
Regression Figure 2.9 EViews Regression Output
Parameters

2.3.3c
Computer Output

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 46
2.3
Estimating the
Regression
Parameters

2.3.4
Other Economic
Models

The simple regression model can be applied to


estimate the parameters of many relationships in
economics, business, and the social sciences
– The applications of regression analysis are
fascinating and useful

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 47
2.4
Assessing the Least Squares Fit

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 48
2.4
Assessing the
Least Squares Fit

We call b1 and b2 the least squares estimators.


– We can investigate the properties of the estimators
b1 and b2 , which are called their sampling
properties, and deal with the following important
questions:
1. If the least squares estimators are random
variables, then what are their expected values,
variances, covariances, and probability
distributions?
2. How do the least squares estimators compare
with other procedures that might be used, and
how can we compare alternative estimators?
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 49
2.4
Assessing the
Least Squares Fit

2.4.1
The Estimator b2 The estimator b2 can be rewritten as:
N
Eq. 2.10 b2   wi yi
i 1

where
xi  x
wi 
 i
Eq. 2.11
( x  x ) 2

It could also be write as:

Eq. 2.12 b2  β2   wi ei

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 50
2.4
Assessing the
Least Squares Fit

2.4.2
The Expected
Values of b1 and b2
We will show that if our model assumptions hold,
then E(b2) = β2, which means that the estimator is
unbiased. We can find the expected value of b2
using the fact that the expected value of a sum is
the sum of the expected values:
E (b2 )  E (b2   wi ei )  E (β 2  w1e1  w2e2  ...  wN eN )
 E (β 2 )  E ( w1e1 )  E ( w2 e2 )  ...  E ( wN eN )
Eq. 2.13  E (β 2 )   E ( wi ei )
 β 2   wi E (ei )
 β2
using E (ei )  0 and E(wi ei )  wi E(ei )
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 51
2.4
Assessing the
Least Squares Fit

2.4.2
The Expected
Values of b1 and b2
The property of unbiasedness is about the average
values of b1 and b2 if many samples of the same size
are drawn from the same population
– If we took the averages of estimates from many
samples, these averages would approach the true
parameter values b1 and b2
– Unbiasedness does not say that an estimate from
any one sample is close to the true parameter value,
and thus we cannot say that an estimate is unbiased
– We can say that the least squares estimation
procedure (or the least squares estimator) is
unbiased
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 52
2.4
Assessing the
Least Squares Fit
Table 2.2 Estimates from 10 Samples

2.4.3
Repeated
Sampling

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 53
2.4
Assessing the
Least Squares Fit
Figure 2.10 Two possible probability density functions for b2

2.4.3
Repeated The variance of b2 is defined as var(b2 )  E[b2  E (b2 )]2
Sampling

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 54
2.4
Assessing the
Least Squares Fit

2.4.4
The Variances and
Covariances of b1
If the regression model assumptions SR1-SR5 are
and b2
correct (assumption SR6 is not required), then the
variances and covariance of b1 and b2 are:

Eq. 2.14 var(b1 )  σ 


2
  x 2
i


 N   xi  x  
2

σ2
var(b2 ) 
  xi  x 
Eq. 2.15 2

 x 
Eq. 2.16 cov(b1 , b2 )  σ  2

   xi  x  
2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 55
2.4
Assessing the MAJOR POINTS ABOUT THE VARIANCES AND COVARIANCES
Least Squares Fit OF b1 AND b2

2.4.4
The Variances and 1. The larger the variance term σ2 , the greater the uncertainty
Covariances of b1
and b2
there is in the statistical model, and the larger the variances
and covariance of the least squares estimators.

The larger the sum of squares,  xi  x  , the smaller the


2
2.
variances of the least squares estimators and the more
precisely we can estimate the unknown parameters.

3. The larger the sample size N, the smaller the variances and
covariance of the least squares estimators.

The larger the term  x , the larger the variance of the least
2
4. i

squares estimator b1.

5. The absolute magnitude of the covariance increases the


larger in magnitude is the sample mean x , and the
covariance has a sign opposite to that of x.
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 56
2.4 Figure 2.11 The influence of variation in the explanatory variable x on
Assessing the precision of estimation (a) Low x variation, low precision (b) High x
Least Squares Fit
variation, high precision
2.4.4
The Variances and
Covariances of b1
The variance of b2 is defined as var(b2 )  Eb2  E (b2 )
2
and b2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 57
2.5
The Gauss-Markov Theorem

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 58
2.5
The Gauss-Markov
Theorem
GAUSS-MARKOV THEOREM

Under the assumptions SR1-SR5 of the linear


regression model, the estimators b1 and b2 have the
smallest variance of all linear and unbiased
estimators of b1 and b2. They are the Best Linear
Unbiased Estimators (BLUE) of b1 and b2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 59
2.5
The Gauss-Markov
Theorem MAJOR POINTS ABOUT THE GAUSS-MARKOV THEOREM

1. The estimators b1 and b2 are “best” when compared to similar


estimators, those which are linear and unbiased. The Theorem does
not say that b1 and b2 are the best of all possible estimators.

2. The estimators b1 and b2 are best within their class because they
have the minimum variance. When comparing two linear and
unbiased estimators, we always want to use the one with the
smaller variance, since that estimation rule gives us the higher
probability of obtaining an estimate that is close to the true
parameter value.

3. In order for the Gauss-Markov Theorem to hold, assumptions SR1-


SR5 must be true. If any of these assumptions are not true, then b1
and b2 are not the best linear unbiased estimators of β1 and β2.

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 60
2.5
The Gauss-Markov
Theorem MAJOR POINTS ABOUT THE GAUSS-MARKOV THEOREM

4. The Gauss-Markov Theorem does not depend on the assumption


of normality (assumption SR6).

5. In the simple linear regression model, if we want to use a linear


and unbiased estimator, then we have to do no more searching.
The estimators b1 and b2 are the ones to use. This explains why
we are studying these estimators and why they are so widely used
in research, not only in economics but in all social and physical
sciences as well.

6. The Gauss-Markov theorem applies to the least squares


estimators. It does not apply to the least squares estimates from a
single sample.

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 61
2.6
The Probability Distributions of the
Least Squares Estimators

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 62
2.6
The Probability
Distributions of the
Least Squares
Estimators

If we make the normality assumption (assumption


SR6 about the error term) then the least squares
estimators are normally distributed:

 σ 2  xi2 
Eq. 2.17 b1 ~ N  β1 , 
 N x  x   2
 i 

 σ2 
Eq. 2.18 b2 ~ N  β 2 , 
  x  x  
2
 i 

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 63
2.6
The Probability
Distributions of the
Least Squares
A CENTRAL LIMIT THEOREM
Estimators

If assumptions SR1-SR5 hold, and if the sample


size N is sufficiently large, then the least squares
estimators have a distribution that approximates the
normal distributions shown in Eq. 2.17 and Eq. 2.18

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 64
2.7
Estimating the Variance of the Error
Term

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 65
2.7
Estimating the
Variance of the
Error Term

The variance of the random error ei is:

var(ei )  σ2  E[ei  E (ei )]2  E (ei )2

if the assumption E(ei) = 0 is correct.

Since the “expectation” is an average value we might


consider estimating σ2 as the average of the
squared errors:

σ̂ 2 
i
e 2

N
where the error terms are ei  yi  β1  β2 xi

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 66
2.7
Estimating the
Variance of the
Error Term

The least squares residuals are obtained by replacing


the unknown parameters by their least squares
estimates:
eˆi  yi  yˆi  yi  b1  b2 xi

σ2 
i
ˆ
e 2

N
There is a simple modification that produces an
unbiased estimator, and that is:

Eq. 2.19 ˆ 2  i
e 2

N 2

 
so that:
E σ̂ 2  σ 2
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 67
2.7
Estimating the
Variance of the
Error Term

2.7.1
Estimating the
Variance and
Replace the unknown error variance σ2 in Eq. 2.14
Covariance of the
Least Squares – Eq. 2.16 by ˆ 2 to obtain:
Estimators

Eq. 2.20 var(b1 )  σˆ 


2
  xi
2 

 N   xi  x  
2

σ̂ 2
var(b2 ) 
  xi  x 
Eq. 2.21 2

 x 
Eq. 2.22 cov(b1 , b2 )  σˆ  2

   xi  x  
2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 68
2.7
Estimating the
Variance of the
Error Term

2.7.1
Estimating the
Variance and
Covariance of the
Least Squares
Estimators
The square roots of the estimated variances are the
“standard errors” of b1 and b2:

Eq. 2.23 se(b1 )  var(b1 )

Eq. 2.24 se(b2 )  var(b2 )

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 69
2.7
Estimating the
Variance of the Table 2.3 Least Squares Residuals
Error Term

2.7.2
Calculations for the
Food Expenditure
Data

σ̂ 2 
i
ˆ
e 2


304505.2
 8013.29
N 2 38

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 70
2.7
Estimating the
Variance of the
Error Term

2.7.2
Calculations for the
Food Expenditure
Data
The estimated variances and covariances for a
regression are arrayed in a rectangular array, or
matrix, with variances on the diagonal and
covariances in the “off-diagonal” positions.

 var(b1 ) cov(b1 , b2 ) 
 
cov(b1 , b2 ) var(b2 ) 

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 71
2.7
Estimating the
Variance of the
Error Term

2.7.2
Calculations for the
Food Expenditure
Data

For the food expenditure data the estimated


covariance matrix is:

C Income
C 1884.442 -85.90316
Income -85.90316 4.381752

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 72
2.7
Estimating the
Variance of the
Error Term

2.7.3
Interpreting the
Standard Errors
The standard errors of b1 and b2 are measures of
the sampling variability of the least squares
estimates b1 and b2 in repeated samples.
– The estimators are random variables. As such,
they have probability distributions, means, and
variances.
– In particular, if assumption SR6 holds, and the
random error terms ei are normally distributed,
then:


b2 ~ N β2 , var(b2 )  σ 2
 x  x 
i
2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 73
2.7
Estimating the
Variance of the
Error Term

2.7.3
Interpreting the
Standard Errors
The estimator variance, var(b2), or its square
root, σb2  var  b2  which we might call the true
standard deviation of b2, measures the sampling
variation of the estimates b2
– The bigger  b2 is the more variation in the least
squares estimates b2 we see from sample to
sample. If  b2 is large then the estimates might
change a great deal from sample to sample
– If  b2 is small relative to the parameter b2,we
know that the least squares estimate will fall
near b2 with high probability

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 74
2.7
Estimating the
Variance of the
Error Term

2.7.3
Interpreting the
Standard Errors

The question we address with the standard error is


‘‘How much variation about their means do the
estimates exhibit from sample to sample?’’

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 75
2.7
Estimating the
Variance of the
Error Term

We estimate σ2, and then estimate  b using:


2.7.3
Interpreting the
Standard Errors 2

se(b2 )  var(b2 )
σ̂ 2

  x  x
2
i

– The standard error of b2 is thus an estimate of


what the standard deviation of many estimates
b2 would be in a very large number of samples,
and is an indicator of the width of the pdf of b2
shown in Figure 2.12
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 76
2.7
Estimating the
Variance of the Figure 2.12 The probability density function of the least squares estimator b2.
Error Term

2.7.3
Interpreting the
Standard Errors

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 77
2.8
Estimating Nonlinear Relationships

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 78
2.8
Estimating
Nonlinear THE WORLD IS NOT LINEAR
Relationships

Economic variables are not always related by straight-line


relationships; in fact, many economic relationships are
represented by curved lines, and are said to display curvilinear
forms.

Fortunately, the simple linear regression model y = β1 + β2 + e


is much more flexible than it looks at first glance, because the
variables y and x can be transformations, involving logarithms,
squares, cubes or reciprocals, of the basic economic variables,
or they can be indicator variables that take only the values
zero and one.

Including these possibilities means the simple linear


regression model can be used to account for nonlinear
relationships between variables
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 79
2.8
Estimating
Nonlinear
Relationships

Consider the linear model of house prices:


Eq. 2.25 PRICE  β1  β2 SQFT  e

where SQFT is the square footage.


– It may be reasonable to assume that larger and
more expensive homes have a higher value for
an additional square foot of living area than
smaller, less expensive, homes

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 80
2.8
Estimating
Nonlinear
Relationships

We can build this into our model in two ways:


1. a quadratic equation in which the explanatory
variable is SQFT2
2. a loglinear equation in which the dependent
variable is ln(PRICE)
– In each case we will find that the slope of the
relationship between PRICE and SQFT is not
constant, but changes from point to point.

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 81
2.8
Estimating
Nonlinear
Relationships

2.8.1
Quadratic
Functions

The quadratic function y = β1 + β2x2 is a parabola


– The elasticity, or the percentage change in y
given a 1% change in x, is:
  slope  x y
 2bx 2 y

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 82
2.8
Estimating
Nonlinear Figure 2.13 A quadratic function
Relationships

2.8.1
Quadratic
Functions

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 83
2.8
Estimating
Nonlinear
Relationships

2.8.2
Using a Quadratic
Model
A quadratic model for house prices includes the
squared value of SQFT, giving:

Eq. 2.26 PRICE  α1  α2 SQFT 2  e

– The slope is:

Eq. 2.27

ˆ
d PRICE   2αˆ SQFT
2
dSQFT
– If α̂ 2  0 , then larger houses will have larger
slope, and a larger estimated price per
additional square foot

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 84
2.8
Estimating
Nonlinear Figure 2.14 A fitted quadratic relationship
Relationships

2.8.2
Using a Quadratic
Model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 85
2.8
Estimating
Nonlinear
Relationships

2.8.2
Using a Quadratic
Model
For 1080 houses sold in Baton Rouge, LA during
mid-2005, the estimated quadratic equation is:
PRICE  55776.56  0.0154SQFT 2
– The estimated slope is:
slope  2  0.0154  SQFT
– The elasticity is:
SQFT
ˆ  slope 
PRICE
SQFT
  2α 2 SQFT  
ˆ
PRICE

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 86
2.8
Estimating
Nonlinear
Relationships

2.8.2
Using a Quadratic
Model

To compute an estimate we must select values for


SQFT and PRICE
– A common approach is to choose a point on the
fitted relationship
• That is, we choose a value for SQFT and
choose for price the corresponding fitted
value

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 87
2.8
Estimating
Nonlinear
Relationships

2.8.2
Using a Quadratic
Model

For houses of 2000, 4000 and 6000 square feet,


the estimated elasticities are:
1.05 using PRÎCE  $117,461.77
1.63 using PRÎCE  $302,517.39
1.82 using PRÎCE  $610,943.42
respectively

For a 2000-square-foot house, we estimate that a


1% increase in house size will increase price by
1.05%

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 88
2.8
Estimating
Nonlinear
Relationships

2.8.3
A Log-Linear
Function
The log-linear equation ln(y) = a + bx has a
logarithmic term on the left-hand side of the
equation and an untransformed (linear) variable on
the right-hand side
– Both its slope and elasticity change at each
point and are the same sign as b
• The slope is:
dy dx  by
– The elasticity, the percentage change in y given
a 1% increase in x, at a point on this curve is:
  slope  x y  bx
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 89
2.8
Estimating
Nonlinear
Relationships

2.8.3
A Log-Linear
Function

Using the slope expression, we can solve for a


semi-elasticity, which tells us the percentage
change in y given a 1-unit increase in x:
100  dy dx 
Eq. 2.28 η  100b
dx

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 90
2.8
Estimating
Nonlinear
Relationships

2.8.4
Using a Log-Linear
Model

Consider again the model for the price of a house


as a function of the square footage, but now
written in semi-log form:
Eq. 2.29 ln  PRICE   γ1  γ 2 SQFT  e
– This logarithmic transformation can regularize
data that is skewed with a long tail to the right

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 91
2.8
Estimating
Nonlinear Figure 2.16 (a) Histogram of PRICE (b) Histogram of ln(PRICE)
Relationships

2.8.4
Using a Log-Linear
Model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 92
2.8
Estimating
Nonlinear
Relationships

2.8.4
Using a Log-Linear
Model

Using the Baton Rouge data, the fitted log-linear


model is:
 
ln PRICE  10.8386  0.0004113SQFT
– To obtain predicted price take the anti-
logarithm, which is the exponential function:

 
PRICE  exp ln PRICE   exp 10.8386  0.0004113SQFT 
 

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 93
2.8
Estimating
Nonlinear Figure 2.17 The fitted log-linear model
Relationships

2.8.4
Using a Log-Linear
Model

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 94
2.8
Estimating
Nonlinear
Relationships

2.8.4
Using a Log-Linear
Model

The slope of the log-linear model is:



d PRICE   γ̂ PRICE  0.0004113PRICE
2
dSQFT
For a house with a predicted PRICE of $100,000,
the estimated increase in PRICE for an additional
square foot of house area is $41.13, and for a
house with a predicted PRICE of $500,000, the
estimated increase in PRICE for an additional
square foot of house area is $205.63

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 95
2.8
Estimating
Nonlinear
Relationships

2.8.4
Using a Log-Linear
Model
The estimated elasticity is:
ˆ  γ̂2 SQFT  0.0004113SQFT
– For a house with 2000-square-feet, the
estimated elasticity is 0.823:
• A 1% increase in house size is estimated to
increase selling price by 0.823%
– For a house with 4000 square feet, the
estimated elasticity is 1.645:
• A 1% increase in house size is estimated to
increase selling price by 1.645%

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 96
2.8
Estimating
Nonlinear
Relationships

2.8.4
Using a Log-Linear
Model

Using the ‘‘semi-elasticity’’ defined in Eq. 2.28 we


can say that, for a one-square-foot increase in size,
we estimate a price increase of 0.04%
– Or, perhaps more usefully, we estimate that a
100-square-foot increase will increase price by
approximately 4%.

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 97
2.8
Estimating
Nonlinear
Relationships

2.8.5
Choosing a
Functional Form

We should do our best to choose a functional form


that is:
– consistent with economic theory
– that fits the data well
– that is such that the assumptions of the
regression model are satisfied

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 98
2.8
Estimating
Nonlinear
Relationships

2.8.5
Choosing a
Functional Form

In real-world problems it is sometimes difficult to


achieve all these goals
– Furthermore, we will never truly know the
correct functional relationship, no matter how
many years we study econometrics
– The truth is out there, but we will never know it
– In applications of econometrics we must simply
do the best we can to choose a satisfactory
functional form

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 99
2.9
Regression with Indicator Variables

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 100
2.9
Regression with
Indicator Variables

An indicator variable is a binary variable that takes the


values zero or one; it is used to represent a
nonquantitative characteristic, such as gender, race, or
location
1 house is in University Town
UTOWN  
0 house is in Golden Oaks

PRICE  β1  β2UTOWN  e

– How do we model this?

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 101
2.9
Regression with
Indicator Variables Figure 2.18 Distributions of house prices

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 102
2.9
Regression with
Indicator Variables

When an indicator variable is used in a regression,


it is important to write out the regression function
for the different values of the indicator variable
β1β1 
β 2β 2 UTOWN 1 1
ififUTOWN
 PRICE   
EE PRICE
β1β1 UTOWN 0 0
ififUTOWN
– The estimated regression is:
PRICE b b1bbUTOWN
PRICE UTOWN
1 2 2

 215.7325
215.7325 61.5091
61.5091UTOWN
UTOWN
 277.2416
277.2416 UTOWN 1 1
ififUTOWN
  215.7325 UTOWN 0 0

215.7325 ififUTOWN
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 103
2.9
Regression with
Indicator Variables

The least squares estimators b1 and b2 in this


indicator variable regression can be shown to be:

b1  PRICE Golden Oaks


b2  PRICE UniversityTown  PRICE Golden Oaks

– In the simple regression model, an indicator


variable on the right-hand side gives us a way
to estimate the differences between population
means

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 104
Key Words

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 105
Keywords

 assumptions  homoskedastic  regression model


 asymptotic  independent  regression
 B.L.U.E. variable parameters
 biased estimator  least squares  repeated sampling
 degrees of freedom estimates  sampling precision
 dependent variable  least squares  sampling properties
 deviation from the estimators  scatter diagram
mean form  least squares  simple linear
 econometric model principle regression function
 economic model  least squares  specification error
 elasticity residuals  unbiased estimator
 Gauss-Markov  linear estimator
Theorem  prediction
 heteroskedastic  random error term

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 106
Appendices

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 107
 2A Derivation of the Least Squares Estimates
 2B Deviation from the Mean Form of b2
 2C b2 is a Linear Estimator
 2D Derivation of Theoretical Expression for b2
 2E Deriving the Variance of b2
 2F Proof of the Gauss-Markov Theorem

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 108
2A
Derivation of the
Least Squares
Estimates

Eq. 2A.1 S (β1 ,β 2 )   ( yi  β1  β 2 xi ) 2


i 1

S
 2 Nβ1  2 yi  2   xi  β 2
Eq. 2A.2 β1
S
 2   xi2  β 2  2 xi yi  2   xi β1
β 2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 109
2A
Derivation of the
Least Squares Figure 2A.1 The sum of squares function and the minimizing values b 1 and b2
Estimates

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 110
2A
Derivation of the
Least Squares
Estimates

Set the derivatives equal to zero to get two


equations:
 
2  yi  Nb1   xi b2  0
2 x y   x b   x b   0
i i i 1
2
i 2

Simplify these to:


Eq. 2A.3 Nb1    xi  b2   yi

Eq. 2A.4
 x b   x b   x y
i 1
2
i 2 i i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 111
2A
Derivation of the
Least Squares
Estimates

Solving the two equations simultaneously, we


get for b2:

N  xi yi   xi  yi
b2 
N  x   xi 
Eq. 2A.5 2
2
i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 112
2B
Deviation from the
Mean Form of b2

We can rewrite the equation 2A.5 by noting


the following:

 i x  x 2   i
x 2
 2 x  i
x  N x 2

 1 
  xi2  2 x  N  xi   Nx 2
Eq. 2B.1  N 
  xi2  2 Nx 2  Nx 2
  xi2  Nx 2
Also note that:

 ix  x 2
  i
x 2
 N x 2

  xi2  x  xi
Eq. 2B.2

x 
 x 2 i
2

i
N
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 113
2B
Deviation from the
Mean Form of b2

Finally, we have:

 x  x  y  y    x y
i i i i  Nx y
Eq. 2B.3
  xi yi 
 x y i i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 114
2B
Deviation from the
Mean Form of b2

We can rewrite b2 in deviation from the mean


form as:

 x  x  y  y 
b2  i i

 x  x  i
2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 115
2B
Deviation from the
Mean Form of b2

First note that we will always have:

 x  x   0
i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 116
2C
b2 is a Linear
Estimator

Now rewrite our formula for b2 and use this fact:


  x  x  y  y 
b2  i i

 x  x 
2
i

 x  x  y  y x  x
 i i i

 x  x 
2
i

 x  x y
 i i

 x  x 
2
i

  xi  x  
  2  i
y
   xi  x  
 
  wi yi
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 117
2D
Derivation of
Theoretical
Expression for b2

To obtain Eq. 2.12, replace yi in Eq. 2.11 by


yi = β1+ β2xi +ei and simplify:

b2  w y i i

  w β  β x  e 
i 1 2 i i

 wβ β w x  we
i 1 2 i i i i

β we
2 i i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 118
2D
Derivation of
Theoretical
Expression for b2

For this, we used the facts that:


   xi  x  
 wi    x  x 2 
 i  
1
2  i
 x  x
  xi  x 
0
w x 1 i i

β w x β
2 i i 2

x  x   0
i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 119
2D
Derivation of
Theoretical
Expression for b2

We can show that  wi xi  1 by using:

 x  x     xi  x  xi  x 
2
i

   xi  x xi  x   xi  x 
   xi  x xi

 x  x x
so that:  w x  x  x  i i


i i 2
i

 x  x x
 i i

 x  x x i i

1
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 120
2E
Deriving the
Variance of b2

First note that:

b2  β2   wi ei

and that:

var(b2 )  Eb2  E (b2 )


2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 121
2E
Deriving the
Variance of b2

var(b2 )  E β 2   wi ei  β 2 
2

 E   wi ei 
2

 
 E   wi ei  2 wi w j ei e j 
2 2

 i j 
 
  wi2 E ei2  2 wi w j E  ei e j 
i j

 σ 2  wi2
σ2

x  x 
2
i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 122
2E
Deriving the
Variance of b2

σ  var(ei )  E  ei  E (ei )   E ei  0  E ei2


2 2 2
 


cov  ei , e j   E  ei  E  ei   e j  E  e j    E  ei e j   0
  
 
 xi  x   x  x   1
2 2
 
 wi   
2
2
 i

  xi  x     x  x     x  x 2 2
2 2
  i
  i

var(aX  by )  a 2 var( X )  b 2 var(Y )  2 cov( X , Y )


Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 123
2E
Deriving the
Variance of b2

var(b2 )  var   2   wi ei  [since β 2 is a constant]

  wi2 var  ei    wi w j cov(ei , e j )


i j
[generalizing the variance rule]

  wi2 var(ei ) [using cov(ei,ej) = 0]

 σ 2  wi2 [using var(ei) = 2 ]

σ2

x  x 
2
i

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 124
2F
Proof of the
Gauss-Markov
Theorem

To begin, let b2*   ki yi be any other linear


estimator of β1. Also, suppose ki = wi + ci.
Then:

b2*   ki yi    wi  ci  yi    wi  ci  β1  β 2 xi  ei 
   wi  ci  β1    wi  ci  β 2 xi    wi  ci  ei
Eq. 2F.1
 β1  wi  β1  ci  β 2  wi xi  β 2  ci xi    wi  ci  ei
 β1  ci  β 2  β 2  ci xi    wi  ci  ei

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 125
2F
Proof of the
Gauss-Markov
Theorem

Now:

Eq. 2F.2
 
E b2*  β1  ci  β 2  β 2  ci xi    wi  ci  E  ei 
 β1  ci  β 2  β 2  ci xi

Eq. 2F.3
ci  0 and c x i i 0

Eq. 2F.4 b2*   ki yi  β 2    wi  ci  ei

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 126
2F
Proof of the
Gauss-Markov
Theorem

For now, observe that :


 ci  xi  x   1 x
i i 
c w     2  i i
c x  2  i
c 0
   xi  x     xi  x    xi  x 
2

Now we can write:


var(b2* )  var β 2    wi  ci  ei 

   wi  ci  var  ei 
2

w  c 
2
σ 2
i i

 σ w  σ c
2 2
i
2 2
i

 var  b   σ  c
2
2 2
i

 var  b2 
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 127
2G
Monte Carlo
Simulation

Monte Carlo simulation experiments use random


number generators to replicate the random way that
data are obtained
– In Monte Carlo simulations we specify a data
generation process and create samples of artificial
data
• Then we ‘‘try out’’ estimation methods on the
data we have created
– We create many samples of size N and examine the
repeated sampling properties of the estimators
• In this way, we can study how statistical
procedures behave under ideal, as well as not so
ideal, conditions
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 128
2G
Monte Carlo
Simulation

The data generation process for the simple linear


regression model is given by:
yi  E  yi | xi   ei
 b1  b2 xi  ei
i  1,..., N
– Each value of the dependent variable yi is
obtained, or generated, by adding a random
error ei to the regression function E(yi |xi )
– To simulate values of yi we create values for the
systematic portion of the regression relationship
E(yi |xi ) and add to it the random error ei
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 129
2G
Monte Carlo
Simulation

2G.1
The Regression
Function

To create the variables for the regression function,


we must:
1. Select sample size N
2. Choose xi values
3. Choose β1 and β2

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 130
2G
Monte Carlo
Simulation

2G.2
The Random Error To be consistent with assumptions SR2–SR4 the
random errors should have mean zero, a constant
variance and be uncorrelated with one another
– We can generate random numbers
– Of course the computer-generated numbers
cannot be truly random, because they are
generated by a computer code
– The random numbers created by computer
software are ‘‘pseudorandom,’’ in that they
behave like random numbers

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 131
2G
Monte Carlo
Simulation Figure 2G.1 The true probability density functions of the data

2G.3
Theoretically True
Values

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 132
2G
Monte Carlo
Simulation Figure 2G.2 The true probability density functions of the estimator b2

2G.3
Theoretically True
Values

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 133
2G
Monte Carlo
Simulation

2G.4
Creating a Sample
of Data

yˆi  75.7679  11.9683 xi


 se  25.7928 1.6313

σˆ  51.5857
Variance-Covariance Matrix
b1 b2
b1 665.2699 -39.9162
b2 -39.9162 2.6611

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 134
2G
Monte Carlo
Simulation

2G.5
Monte Carlo
Objectives
What do we hope to achieve with a Monte Carlo
experiment?
– We would like to verify that under SR1–SR5
the least squares estimators are unbiased
– We would like to verify that under SR1–SR5
the least squares estimators have sampling
variances given by Eq. 2.14 and Eq. 2.16
– We would like to verify that the estimator of the
error variance Eq. 2.19 is unbiased
– Because we have assumed the random errors
are normal, SR6, we expect the least squares
estimates to have a normal distribution.
Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 135
2G
Monte Carlo
Simulation Table 2G.1 Summary of 1,000 Monte Carlo Samples

2G.6
Monte Carlo
Results

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 136
2G
Monte Carlo
Simulation Figure 2G.3 The sampling distribution of b2 in 1000 Monte Carlo samples

2G.6
Monte Carlo
Results

Principles of Econometrics, 4th Edition Chapter 2: The Simple Linear Regression Model Page 137

You might also like