Professional Documents
Culture Documents
Advanced Quantitative Research Methodology, Lecture Notes:: Gary King
Advanced Quantitative Research Methodology, Lecture Notes:: Gary King
Gary King
http://GKing.Harvard.edu
February 2, 2014
1
Copyright
c 2014 Gary King, All Rights Reserved.
Gary King (Harvard) The Basics February 2, 2014 1 / 61
Who Takes This Course?
Most Gov Dept grad students doing empirical work, the 2nd course in
their methods sequence (Gov2001)
Most Gov Dept grad students doing empirical work, the 2nd course in
their methods sequence (Gov2001)
Grad students from other departments (Gov2001)
Most Gov Dept grad students doing empirical work, the 2nd course in
their methods sequence (Gov2001)
Grad students from other departments (Gov2001)
Undergrads (Gov1002)
Most Gov Dept grad students doing empirical work, the 2nd course in
their methods sequence (Gov2001)
Grad students from other departments (Gov2001)
Undergrads (Gov1002)
Non-Harvard students, visitors, faculty, & others (online through the
Harvard Extension school, E-2001)
Most Gov Dept grad students doing empirical work, the 2nd course in
their methods sequence (Gov2001)
Grad students from other departments (Gov2001)
Undergrads (Gov1002)
Non-Harvard students, visitors, faculty, & others (online through the
Harvard Extension school, E-2001)
Some of the best experiences here: getting to know people in other
fields
All math requires two parts: proof and concepts & intuition
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
Use rigorous statistical theory — when needed
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
Use rigorous statistical theory — when needed
Insure we understand the intuition — always
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
Use rigorous statistical theory — when needed
Insure we understand the intuition — always
Always traverse from theoretical foundations to practical applications
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
Use rigorous statistical theory — when needed
Insure we understand the intuition — always
Always traverse from theoretical foundations to practical applications
Fewer proofs, more concepts, better practical knowledge
All math requires two parts: proof and concepts & intuition
Different classes emphasize:
Baby Stats: dumbed down proofs, vague intuition
Math Stats: rigorous mathematical proofs
Practical Stats: deep concepts and intuition, proofs when needed
Goal: how to do empirical research, in depth
Use rigorous statistical theory — when needed
Insure we understand the intuition — always
Always traverse from theoretical foundations to practical applications
Fewer proofs, more concepts, better practical knowledge
Do you have the background for this class? A Test: What’s this?
b = (X 0 X )−1 X 0 y
j.mp/G2001
j.mp/G2001
The syllabus gives topics, not a weekly plan.
j.mp/G2001
The syllabus gives topics, not a weekly plan.
We will go as fast as possible subject to everyone following along
j.mp/G2001
The syllabus gives topics, not a weekly plan.
We will go as fast as possible subject to everyone following along
We cover different amounts of material each week
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career,
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
Instead, we teach you the fundamentals, the underlying theory of
inference, from which statistical models are developed:
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
Instead, we teach you the fundamentals, the underlying theory of
inference, from which statistical models are developed:
We will reinvent existing methods by creating them from scratch.
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
Instead, we teach you the fundamentals, the underlying theory of
inference, from which statistical models are developed:
We will reinvent existing methods by creating them from scratch.
We will learn: its easy to invent new methods too, when needed.
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
Instead, we teach you the fundamentals, the underlying theory of
inference, from which statistical models are developed:
We will reinvent existing methods by creating them from scratch.
We will learn: its easy to invent new methods too, when needed.
The fundamentals help us pick up new methods created by others.
We could teach you the latest and greatest methods, but when you
graduate they will be old
We could teach you all the methods that might prove useful during
your career, but when you graduate you will be old
Instead, we teach you the fundamentals, the underlying theory of
inference, from which statistical models are developed:
We will reinvent existing methods by creating them from scratch.
We will learn: its easy to invent new methods too, when needed.
The fundamentals help us pick up new methods created by others.
This helps us separate the conventions from underlying statistical
theory. (How to get an F in Econometrics: follow advice from
Psychometrics. Works in reverse too, even when the foundations are
identical.)
Standard version
Standard version
Yi = xi β + i = systematic + stochastic
Standard version
Yi = xi β + i = systematic + stochastic
i ∼ fN (0, σ 2 )
Standard version
Yi = xi β + i = systematic + stochastic
i ∼ fN (0, σ 2 )
Alternative version
Standard version
Yi = xi β + i = systematic + stochastic
i ∼ fN (0, σ 2 )
Alternative version
Yi ∼ fN (µi , σ 2 ) stochastic
Standard version
Yi = xi β + i = systematic + stochastic
i ∼ fN (0, σ 2 )
Alternative version
Yi ∼ fN (µi , σ 2 ) stochastic
µi = xi β systematic
Yi ∼ f (θi , α) stochastic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
θi a systematic feature of the density that varies over i
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
θi a systematic feature of the density that varies over i
α ancillary parameter (feature of the density constant over i)
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
θi a systematic feature of the density that varies over i
α ancillary parameter (feature of the density constant over i)
g (·) functional form
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
θi a systematic feature of the density that varies over i
α ancillary parameter (feature of the density constant over i)
g (·) functional form
Xi explanatory variables
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
where
Yi random outcome variable
f (·) probability density
θi a systematic feature of the density that varies over i
α ancillary parameter (feature of the density constant over i)
g (·) functional form
Xi explanatory variables
β effect parameters
Yi ∼ f (θi , α) stochastic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
Yi ∼ f (θi , α) stochastic
θi = g (Xi , β) systematic
We (ultimately) will
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
the chosen family (model dependence)
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
the chosen family (model dependence)
If the true relationship falls outside the assumed class, we
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
the chosen family (model dependence)
If the true relationship falls outside the assumed class, we
Have specification error, and potentially bias
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
the chosen family (model dependence)
If the true relationship falls outside the assumed class, we
Have specification error, and potentially bias
Still get the best [linear,logit,etc] approximation to the correct
functional form.
We (ultimately) will
Assume a class of functional forms (each form is flexible and maps out
many potential relationships)
Within the class, choose a member of the class by estimating β
Since data contain (sampling, measurement, random) error, we will be
uncertain about:
the member of the chosen family (sampling error)
the chosen family (model dependence)
If the true relationship falls outside the assumed class, we
Have specification error, and potentially bias
Still get the best [linear,logit,etc] approximation to the correct
functional form.
May be close or far from the truth:
Rb
For both: Pr(a ≤ Y ≤ b) = a P(Y )dY
Rb
For both: Pr(a ≤ Y ≤ b) = a P(Y )dY
For discrete: Pr(y ) = P(y )
Rb
For both: Pr(a ≤ Y ≤ b) = a P(Y )dY
For discrete: Pr(y ) = P(y )
For continuous: Pr(y ) = 0 (why?)
Gary King (Harvard) The Basics 31 / 61
What you should know about every pdf
Expected value:
Expected value:
X
E (Y ) = y P(y )
all y
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Variance:
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Variance:
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Variance:
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Variance:
Expected value:
X
E (Y ) = y P(y )
all y
= 0 Pr(0) + 1 Pr(1)
=π
Variance:
How do we compute E (Y 2 )?
Gary King (Harvard) The Basics 36 / 61
Expected values of functions of random variables
X
E [g (Y )] = g (y )P(y )
all y
X
E [g (Y )] = g (y )P(y )
all y
or
X
E [g (Y )] = g (y )P(y )
all y
or
Z ∞
E [g (Y )] = g (y )P(y )
−∞
X
E [g (Y )] = g (y )P(y )
all y
or
Z ∞
E [g (Y )] = g (y )P(y )
−∞
For example,
X
E [g (Y )] = g (y )P(y )
all y
or
Z ∞
E [g (Y )] = g (y )P(y )
−∞
For example,
X
E (Y 2 ) = y 2 P(y )
all y
X
E [g (Y )] = g (y )P(y )
all y
or
Z ∞
E [g (Y )] = g (y )P(y )
−∞
For example,
X
E (Y 2 ) = y 2 P(y )
all y
= 0 Pr(0) + 12 Pr(1)
2
X
E [g (Y )] = g (y )P(y )
all y
or
Z ∞
E [g (Y )] = g (y )P(y )
−∞
For example,
X
E (Y 2 ) = y 2 P(y )
all y
= 0 Pr(0) + 12 Pr(1)
2
=π
0 0 0 1 0 0 1 1 0 0 1 1 1 0 ...
0 0 0 1 0 0 1 1 0 0 1 1 1 0 ...
Explanation:
Explanation:
N
y because (1 0 1) and (1 1 0) are both y = 2.
Explanation:
N
y because (1 0 1) and (1 1 0) are both y = 2.
π y because y successes with π probability each (product taken due to
independence)
Explanation:
N
y because (1 0 1) and (1 1 0) are both y = 2.
π y because y successes with π probability each (product taken due to
independence)
(1 − π)N−y because N − y failures with 1 − π probability each
Explanation:
N
y because (1 0 1) and (1 1 0) are both y = 2.
π y because y successes with π probability each (product taken due to
independence)
(1 − π)N−y because N − y failures with 1 − π probability each
Mean E (Y ) = Nπ
Explanation:
N
y because (1 0 1) and (1 1 0) are both y = 2.
π y because y successes with π probability each (product taken due to
independence)
(1 − π)N−y because N − y failures with 1 − π probability each
Mean E (Y ) = Nπ
Variance V (Y ) = π(1 − π)/N.
Gary King (Harvard) The Basics 40 / 61
How to simulate from the Binomial distribution
Yi ∼ N(yi |µi , Σ)
Yi ∼ N(yi |µi , Σ)
Yi ∼ N(yi |µi , Σ)
Yi ∼ N(yi |µi , Σ)
Mathematical form:
Yi ∼ N(yi |µi , Σ)
Mathematical form:
−k/2 −1/2 1
N(yi |µi , Σ) = (2π) |Σ| exp − (yi − µi )0 Σ−1 (yi − µi )
2
Yi ∼ N(yi |µi , Σ)
Mathematical form:
−k/2 −1/2 1
N(yi |µi , Σ) = (2π) |Σ| exp − (yi − µi )0 Σ−1 (yi − µi )
2
6
4
4
2
2
0
0
1 1
0.8 0.8
0.6 1 1 1
0.6
0.8 0.8 0.8
0.4 0.4
0.6 0.6 1
βwi βwi
0.6
0.2 0.4 0.2 0.4 0.8
0
0.2 βbi 0
0.2 βbi βwi
0.4 0.6
0 0.2 0.4
βbi
0
0.2
0 0
(a) 0.5 0.5 0.15 0.15 0 (b) 0.1 0.9 0.15 0.15 0 (c) 0.8 0.8 0.6 0.6 0.5
We will stop here this year and skip to the next set of slides.
Please refer to the slides below for further information on probability
densities and random number generation; they offer more sophisticated .
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Intuitive?
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
Γ(α + β) α−1
Beta(y |α, β) = y (1 − y )β−1
Γ(α)Γ(β)
where, Γ(x) is the gamma function:
Z ∞
Γ(x) = z x−1 e −z dz
0
α
Set µ = E (Y ) = (α+β)
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
,
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
, solve for α
and β and substitute in.
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
, solve for α
and β and substitute in.
Result:
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
, solve for α
and β and substitute in.
Result:
Γ µγ −1 + (1 − µ)γ −1 µγ −1 −1
−1
beta(y |µ, γ) = −1 −1
y (1 − y )(1−µ)γ −1
Γ (µγ ) Γ [(1 − µ)γ ]
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
, solve for α
and β and substitute in.
Result:
Γ µγ −1 + (1 − µ)γ −1 µγ −1 −1
−1
beta(y |µ, γ) = −1 −1
y (1 − y )(1−µ)γ −1
Γ (µγ ) Γ [(1 − µ)γ ]
α µ(1−µ)γ αβ
Set µ = E (Y ) = (α+β) and (1+γ) = V (Y ) = (α+β)2 (α+β+1)
, solve for α
and β and substitute in.
Result:
Γ µγ −1 + (1 − µ)γ −1 µγ −1 −1
−1
beta(y |µ, γ) = −1 −1
y (1 − y )(1−µ)γ −1
Γ (µγ ) Γ [(1 − µ)γ ]
How to simulate
How to simulate
How to simulate
How to simulate
How to simulate
How to simulate
How to simulate
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B) Pr (B)
Pr(B)
Plan:
All assumptions are about the events that occur between the start
and when we observe the count. The process of event generation is
assumed not observed.
All assumptions are about the events that occur between the start
and when we observe the count. The process of event generation is
assumed not observed.
0 events occur at the start of the period
All assumptions are about the events that occur between the start
and when we observe the count. The process of event generation is
assumed not observed.
0 events occur at the start of the period
Only observe number of events at the end of the period
All assumptions are about the events that occur between the start
and when we observe the count. The process of event generation is
assumed not observed.
0 events occur at the start of the period
Only observe number of events at the end of the period
No 2 events can occur at the same time
All assumptions are about the events that occur between the start
and when we observe the count. The process of event generation is
assumed not observed.
0 events occur at the start of the period
Only observe number of events at the end of the period
No 2 events can occur at the same time
Pr(event at time t | all events up to time t − 1) is constant for all t.
e −λ λyi
(
yi ! for yi = 0, 1, . . .
Poisson(y |λ) =
0 otherwise
e −λ λyi
(
yi ! for yi = 0, 1, . . .
Poisson(y |λ) =
0 otherwise
E (Y ) = λ
e −λ λyi
(
yi ! for yi = 0, 1, . . .
Poisson(y |λ) =
0 otherwise
E (Y ) = λ
V (Y ) = λ
e −λ λyi
(
yi ! for yi = 0, 1, . . .
Poisson(y |λ) =
0 otherwise
E (Y ) = λ
V (Y ) = λ
That the variance goes up with the mean makes sense, but should they
be equal?
e −λ λyi
(
yi ! for yi = 0, 1, . . .
Poisson(y |λ) =
0 otherwise
E (Y ) = λ
V (Y ) = λ
That the variance goes up with the mean makes sense, but should they
be equal?
Recall:
Recall:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B)Pr(B)
Pr(B)
Recall:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B)Pr(B)
Pr(B)
Z ∞
2
NegBin(y |φ, σ ) = Poisson(y |λ) × gamma(λ|φ, σ 2 )dλ
0
Recall:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B)Pr(B)
Pr(B)
Z ∞
2
NegBin(y |φ, σ ) = Poisson(y |λ) × gamma(λ|φ, σ 2 )dλ
0
Z ∞
= P(y , λ|φ, σ 2 )dλ
0
Recall:
Pr(AB)
Pr(A|B) = =⇒ Pr(AB) = Pr(A|B)Pr(B)
Pr(B)
Z ∞
2
NegBin(y |φ, σ ) = Poisson(y |λ) × gamma(λ|φ, σ 2 )dλ
0
Z ∞
= P(y , λ|φ, σ 2 )dλ
0
Γ σ2φ−1 + yi σ 2 − 1 yi −φ
2 σ2 −1
= σ
y !Γ φ σ2
i σ 2 −1