Professional Documents
Culture Documents
Econometric Lec1
Econometric Lec1
Econometric Lec1
ECONOMETRICS
1
AND
REGRESSION ANALYSIS LECTURE
1
Chapter 0
Introduction
2
What is Econometrics
3
1. Economics Theory
“ Keynes postulated a positive relationship between consumption
and incomes”, i.e., people’s income
2. Mathematical Expression:
Consumption = f(Income) ==> C = f(Y)
MPC = dC/dY = f’(Y) > 0 ;assume 0 < MPC < 1
3. Statistics:
Year C Y Find the mean, variance,
standard deviation,
1980 2447.1 3776.3 correlation, etc.
1981 2476.9 3841.1
…. …. ….
4. Econometric - Regression model
C t = b1 + b2 Y t + u t => DC/DY = b2
=> estimating the relationship 4
The Role of Econometrics
1. Economic theory
2. Economic data
3. Methods of model constructed
5
Economic Relationships:
Stock Market Index money supply
government
budget
Interest rate
Exchange
inflation trade Rate
deficit Properties Market
unemployment
Wage
}
economic theory
economic
decisions
economic data
8
Methodology of Econometrics
• Statement of theory or hypothesis
• Specification of the mathematical model of the
theory
• Specification of the econometric model of the
theory
• Obtaining data for the analysis.
• Estimation with statistical properties.
• Hypothesis testing
• Analyze and evaluate implications of the results
• Forecasting or prediction
• Using the model for control or policy purpose
9
10
Example: The Consumption Function
C = f(inc)
The nature of
Regression
Analysis
13
Regression Analysis
14
Purpose of Regression Analysis
15
Terminology and Notation
Y = b1 + b2 X + u
Dependent Explanatory
Explained Independent
Predictand Predictor
Regressand Regressor
Response Stimulus or control
Endogenous Exogenous
16
Some Examples
• Consumption on Income
• GDP on Export, Money Supply
• Sales on Season, Advertising
Expenditures
• Income on Gender, Education, Years
of Traning
17
Some Comparisons
18
Data
• Time-series Data
• Cross-Sectional Data
• Pooled or Panel Data
19
Time series data
20
Cross-section data and Pool (Panel) data
21
Chapter 2
Two-variable
Regression
Analysis : Some Basic
Ideas
22
Population Regression Function
23
Weekly Food Expenditures
Y = dollars spent each week on food items.
X = consumer’s family weekly income.
DE(Y|X)
DE(Y|X) b2=
DX
Dx slope
b1{
Intercept X (income)
• Linear in variables
• Linear in Parameters
27
Stochastic Specification of PRF
ui = Yi - E(Y|Xi)
or Yi = E(Y|Xi) + u i
Shochastic error or
or Yi = b1 + b2 X + u i Stochastic disturbance
28
The Error Term
• Y is a random variable composed of two parts:
Y^ = 65 = ^b + b^ (80)
1 1 2
Y^ = 65 = ^b + b^ (80)
Estimated average: 2 1 2
Y^3 = 65 = ^b1 + b^2 (80)
Y^4 = 65 = ^b1 + b^2 (80)
Y^5 = 65 = ^b1 + b^2 (80)
30
The reasons for stochastic disturbance
• Vagueness of theory
• Unavailability of data
• Direct effect vs indirect effect
• (Core variables vs peripheral variables)
• Intrinsic randomness in human behaviour
• Poor proxy variables
• Principle of parsimony
• wrong functional form
31
Unobservable Nature of Error Term
{ E(Y|x)=b1+b2x
Y3 (PRF)
}
Y2 ^
^ u2
Y 2
u2
E(Y|x2)
Y1 .} u 1
x
x1 x2 x3 x4
The relationship among Yi, ui and the true regression line.
33
The Sample Regression Function (SRF)
(SRF2)
Y ^ ^ ^
Y = b1 + b2x
Y4 ^u . ^Y = ^b + ^b x
4{ 1 2
(SRF1)
Y3 .}^
u3
Y2 ^u .
2{
^
Y1 } 1
.u
x
x1 x2 x3 x4
34
Different samples will have different SRFs)
SRF:
^ ^ ^
Yi = b1 + b2 Xi
^ ^
or Yi = b1 + b2Xi + u^ i Residual
or Yi = b1 + b2 Xi + ei
PRF:
E(Y|X) = b1 + b2 Xi
Yi = b1 + b2 Xi + u i Error term or
Disturbance
^
Yi = estimator of Yi (E(Y|xi)
^
bi or bi = estimator of bi 35
Chapter 3
Two-variable
Regression Model: The
Problem of Estimation
36
Ordinary Least Squares (OLS) Method
Yi = b1 + b2Xi + ui
u^ i = Y i - b^1 - b^2X i
¶f(.)
= - 2 S (Y i - b^1 - b^2Xi )
¶ b^1
¶f(.)
= - 2 S Xi (Yi - b^1 - b^2Xi )
¶ b^2
Set each of these two derivatives equal to zero and
solve these two equations for the two unknowns: b^1 b^238
To minimize f(.), you set the two
derivatives equal to zero to get:
¶f(.)
= - 2 S (Y i – b1 – b2Xi ) = 0
¶ b^1
¶f(.)
= - 2 S xi (Yi - b1 – b2Xi ) = 0
¶ b^2
39
- 2 S (Y i - b1 – b2Xi ) = 0
-2S Xi (Y i – b1 – b2Xi ) = 0
S Yi - nb1 – b2 SXi = 0
2
S Xi Yi - b1 S X i - b2 S Xi = 0
nb1 + b2 S Xi = S Yi
S Xi + b2S Xi S Xi Yi
2
b1 =
40
n SXi b1
= S Yi
S i S i
X X 2
b2 = S Xi Yi
Solve the two unknowns
n SXi Yi - S Xi SYi
b2 =
n SX i - (SXi )
2 2
b1 = Y - b2 x 41
Y
Y4
. ^
Y = b1 + b2X
^
Y*
^*
Y3
.
{.
^u*
4
^*
^
Y* = b*1 + b* 2X
^
Y*
.
1 . 2
^
u*3{ Y4
u*2 {. Y2
^ .
{
^
u*1
.
Y1
Y3
x1 x2 x3 x4 x
re
d itu
en
p
.
ex .
Y
i
r e
itu
n d
p e
ex
.
.
.
x1 x2 x3 income xt
The variance of Yi increases as family income,
xi, increases. 45
Assumptions (continue)
Y ~ N [(b1+b2X), s2 ]
48
The Error Term Assumptions
49
Simple Linear Regression Model
Yi = b1 + b2 X i + ui
^2
s ^2
s
var(b2) = =
S(xi - x) 2
S 2
xi
b2 is a function of the Yi values but
var(b2) does not involve Yi directly. 53
Variance of b1
Given b1 = y - b2x
the variance of the estimator b1 is:
Sx i
2
Sx i2
var(b1) = s2 = s2
n S(x i - x)
2 2
nSxi
54
Covariance of b1 and b2
-x -x
cov(b1,b2) = s2 = s2
2
S(x t - x) S xi
2
55
Estimating the variance
of the error term, s 2
^u = yi - b1 - b2 xi
i
T
^
S
2
ui
^2
s = i =1
n- 2
^
s is an unbiased estimator of s 2
2
56
Properties of LS Estimators:
Gauss-Markov Theorem
58
The coefficient of Determination, R2 –
A Measure of “goodness of fit”
59
The coefficient of Determination, R2 –
A Measure of “goodness of fit”
^ ^
(Yi - Y) = (Yi - Yi) + (Yi - Y)
u^
To measure variation:
^ 2 ^
S(Yi - Y)2 = S[(Yi - Y) i + (Yi - Y)]2
^ 2 ^
S(Yi - Y)2 = S(Yi - Y) i + S(Yi - Y)2
2 ESS RSS
Define R = =1-
TSS TSS
Su
^ 2
i
=1-
S(Yi - Y)2
^
S(Yi - Y)2
R2 =
S(Yi - Y)2 1 ³ R2 ³ 0
61
Alternative R2 expression
^
ESS S(Yi - Y)2 Sy^ i2
2
R = = =
TSS S(Yi - Y) 2 Sy i
2
^ ^2
S(b2xi )2 b Sx 2
2 i ^ 2 Sxi2
= Sy 2 = = b
i Syi2 2
Syi2
^ 2
2
Sx Sxiyi 2
Sxi2 (Sxiyi)2
= b = =
2 Sy2 Sxi2 Syi2 Sxi2Syi2
62
Explanation of R2
• R2 tells us how well the sample regression
line fits the data
• The value of R2 lets us know how much
percent of variation in the dependent variable
can be explained by the independent
variable or by the regression model.
• Ex: R2=0.8: 80% of the variation in Y can be
explained by X.
63
Y As R2 = 0
SRF
Which SRF ?
Y
As R2 = 1
SRF
65
Numerical Example
• Manual Calculations
• Demonstration of using computer software
• Use the data in table 3.8
66
Chapter 4
Classical Normal Linear
Regression Model
67
Summary of BLUE estimators
Mean
E(b1)=b1 and E(b2)=b2
Variance
Var(b1 )=s 2
Sx i2
and s2
Var(b2)=
nSxi
2
Sxi 2
Se(bk) = var(bk) 68
Estimated Error Variance
T
^
S
2
ui ^ 2) = s2
E(s
^2
s = i =1
n- 2
T
^
S
2
^=
s ^
s =
2 ui
i =1 K
n- 2 # of independent
Variables plus
69 the
Constant term
Normality Assumption for ui
ui ~ N(0,σ2)
Y ~ N [(β1+
β2X), σ2 ]
70
Probability Distribution
of Least Squares Estimators
s2Sx i2
b1 ~ N b1 ,
nSx i2
s2
b2 ~ N b2 ,
Sx i2
71
THE END
72