Professional Documents
Culture Documents
Applied Stochastic Processes: G.A. Pavliotis
Applied Stochastic Processes: G.A. Pavliotis
G.A. Pavliotis
kB T
D= .
6πηa
dX
= F (X ) + Σ(X )ξ(t),
dt
where ξ(t) is a random force.
Study the probability ρ(x, t) of finding a particle at position x at
time t. This probability distribution satisfies the Fokker-Planck
equation:
∂ρ 1
= −∇ · (F (x)ρ) + ∇∇ : (A(x)ρ),
∂t 2
where A(x) = Σ(x)Σ(x)T .
−2
−4
−6
0 5 10 15 20 25 30 35 40 45 50
20
10
−10
−20
−30
−40
−50
0 100 200 300 400 500 600 700 800 900 1000
1
Ztn := √ Snt .
n
0.5
U(t)
0
−0.5
−1
−1.5
0 0.2 0.4 0.6 0.8 1
t
{ω ∈ Ω : X (ω) ∈ A} =: {X ∈ A}
Example
Let I denote a subset of the positive integers. A vector
ρ0 = {ρ0,i , i ∈ I} is a distribution
P on I if it has nonnegative entries and
its total mass equals 1: i∈I ρ0,i = 1.
and Z
P[X ∈ G] = d µX (x), G ∈ B(E ).
G
E(X ) = m (1)
Lemma
Let X be a random variable with characteristic function φ(t) and
assume that it has finite moments. Then
1 (k )
E (X k ) = φ (0).
ik
(c) Zn converges to Z in Lp if
p
lim E Zn − Z = 0.
n→+∞
with x = (x1 , . . . , xk ).
We will say that two processes Xt and Yt are equivalent if they
have same finite dimensional distributions.
From experiments or numerical simulations we can only obtain
information about the (fdd) of a process.
Definition
A stochastic process Xt ∈ L2 is called second-order stationary or
wide-sense stationary if the first moment EXt is a constant and the
second moment E(Xt Xs ) depends only on the difference t − s:
Lemma
Assume that the covariance function C(t) of a second order stationary
process is continuous at t = 0. Then it is continuous for all t ∈ R.
Furthermore, the continuity of C(t) is equivalent to the continuity of the
process Xt in the L2 -sense.
as h → 0.
Assume now that C(t) is continuous. From the above calculation we
have
E|Xt+h − Xt |2 = 2(C(0) − C(h)), (4)
which converges to 0 as h → 0. Conversely, assume that Xt is
L2 -continuous. Then, from the above equation we get
limh→0 C(h) = C(0).
Definition
A function f (x) : R 7→ R is called nonnegative definite if
n
X
f (ti − tj )ci c̄j > 0 (5)
i,j=1
for all n ∈ N, t1 , . . . tn ∈ R, c1 , . . . cn ∈ C.
Proof.
Pn
We will use the notation Xtc := i=1 Xti ci . We have.
n
X n
X
C(ti − tj )ci c̄j = EXti Xtj ci c̄j
i,j=1 i,j=1
Xn n
X
= E Xti ci Xtj c̄j = E Xtc X̄tc
i=1 j=1
= E|Xtc |2 > 0.
Definition
Let Xt be a second order stationary process with covariance C(t)
whose Fourier transform is the measure ρ(dx). The measure ρ(dx) is
called the spectral measure of the process Xt .
The slower the decay of the correlation function, the larger the
correlation time is. We have to assume sufficiently fast decay of
correlations so that the correlation time is finite.
Lemma
Let Y (t) denote the stationary OU process with covariance
function (7). Then the position process (8) is a mean zero Gaussian
process with covariance function
where Z ∞
D= R(t) dt
0
is the diffusion coefficient. Thus, one expects that at sufficiently long
times and under appropriate assumptions on the correlation function,
the time integral of a stationary process will approximate a Brownian
motion with diffusion coefficient D.
This kind of analysis was initiated in G.I. Taylor, Diffusion by
Continuous Movements Proc. London Math. Soc..1922; s2-20:
196-212.
are independent.
If, furthermore, for any t1 , t2 and Borel set B ⊂ R
P(Xt2 +s − Xt1 +s ∈ B)
0.5
U(t)
0
−0.5
−1
−1.5
0 0.2 0.4 0.6 0.8 1
t
Moreover,
E W (t) ⊗ W (s) = min(t, s)I. (12)
1 2
g(x, t) = √ e−x /2t .
2πt
We can easily calculate all moments of the Brownian motion:
Z +∞
1 2
n
E(x (t)) = √ x n e−x /2t dx
2πt −∞
n
1.3 . . . (n − 1)t n/2 , n even,
=
0, n odd.
Lemma
Let W (t) be a standard Brownian motion and consider the process
1 2H
E(WtH WsH ) = s + t 2H − |t − s|2H . (13)
2
Fractional Brownian motion has the following properties.
1
1 When H = 21 , Wt 2 becomes the standard Brownian motion.
2 W0H = 0, EWtH = 0, E(WtH )2 = |t|2H , t > 0.
3 It has stationary increments, E(WtH − WsH )2 = |t − s|2H .
4 It has the following self similarity property
H
(Wαt , t > 0) = (αH WtH , t > 0), α > 0,
Definition
The Poisson process with intensity λ, denoted by N(t), is an
integer-valued, continuous time, stochastic process with independent
increments satisfying
k
e−λ(t−s) λ(t − s)
P[(N(t) − N(s)) = k] = , t > s > 0, k ∈ N.
k!
Both Brownian motion and the Poisson process are
homogeneous (or time-homogeneous): the increments
between successive times s and t depend only on t − s.
Or, Z Z
1 1
sψn (s) ds + t ψn (s) ds = λn ψn (t).
0 t
We differentiate this equation twice:
Z 1
ψn (s) ds = λn ψn′ (t) and − ψn (t) = λn ψn′′ (t),
t
FtX := σ (Xs ; s 6 t) .
Definition
Let {Xt } be a stochastic process defined on a probability space
(Ω, F, µ) with values in E and let FtX be the filtration generated by
{Xt }. Then {Xt } is a Markov process if
IΓ (·) denotes the indicator function of the set Γ. We have also set
E = R.
Theorem
(Ethier and Kurtz 1986, Sec. 4.1) Let P(t, x, Γ) satisfy (20) and
assume that (E , ρ) is a complete separable metric space. Then there
exists a Markov process X in E whose finite-dimensional distributions
are uniquely determined by (21).
Pavliotis (IC) StochProc January 16, 2011 89 / 367
Let Xt be a homogeneous Markov process with initial distribution
ν(Γ) = P(X0 ∈ Γ) and transition function P(x, t, Γ). We can calculate
the probability of finding Xt in a set Γ at time t:
Z
P(Xt ∈ Γ) = P(x, t, Γ)ν(dx).
E
Thus, the initial distribution and the transition function are sufficient to
characterize a homogeneous Markov process. Notice that they do not
provide us with any information about the actual paths of the Markov
process.
P0 = I, Pt+s = Pt ◦ Ps ∀ t, s > 0.
Consequently:
Pt+s = Pt ◦ Ps .
Definition
The operator L : Cb (E ) → Cb (E ) defined above is called the generator
of the Markov process {Xt }.
Pt = exp(Lt).
∂u
= Lu, u(x, 0) = f (x). (23)
∂t
∂γ 1 ∂2γ
= .
∂t 2 ∂x 2
Let µt := Pt∗ µ. This is the law of the Markov process and µ is the
initial distribution. An argument similar to the one used in the
derivation of the backward Kolmogorov equation (23) enables us
to obtain an equation for the evolution of µt :
∂µt
= L ∗ µt , µ0 = µ.
∂t
∂(Pt f )
= L(Pt f ),
∂t
or through the evolution of states
(Schrödinger/Frobenious-Perron)
∂(Pt∗ µ)
= L∗ (Pt∗ µ).
∂t
We can also study Markov processes at the level of trajectories.
We will do this after we define the concept of a stochastic
differential equation.
Definition
A Markov process is called ergodic if the equation
Pt g = g, g ∈ Cb (E ) ∀t > 0
Pt∗ µ = µ.
L∗ µ = 0
L∗ ρ = 0.
Example
Consider a one-dimensional Brownian motion on [0, 1], with periodic
boundary conditions. The generator of this Markov process L is the
d2
differential operator L = 21 dx 2 , equipped with periodic boundary
conditions on [0, 1]. This operator is self-adjoint. The null space of both
L and L∗ comprises constant functions on [0, 1]. Both the backward
Kolmogorov and the Fokker-Planck equation reduce to the heat
equation
∂ρ 1 ∂2ρ
=
∂t 2 ∂x 2
with periodic boundary conditions in [0, 1]. Fourier analysis shows that
the solution converges to a constant at an exponential rate.
d d2
L = −αx + D 2.
dx dx
The null space of L comprises constants in x. Hence, it is an
ergodic Markov process. In order to calculate the invariant
measure we need to solve the stationary Fokker–Planck equation:
where
d d 2h
L∗ h :=
(axh) + D 2 .
dx dx
We can calculate the invariant distribution by solving
equation (26).
The invariant measure of this process is the Gaussian measure
r α
α
µ(dx) = exp − x 2 dx.
2πD 2D
If the initial
Pavliotis (IC) condition of the OU process is distributed
StochProc according
January 16, 2011 112to
/ 367
Let Xt be the 1d OU process and let X0 ∼ N (0, D/α). Then Xt is a
mean zero, Gaussian second order stationary process on [0, ∞)
with correlation function
D −α|t|
R(t) = e
α
and spectral density
D 1
f (x) = .
π x + α2
2
then we can extend the integration over the whole Rd and use
expectations in the definition of the drift and the diffusion coefficient.
Indeed, ,let k = 0, 1, 2 and notice that
Z
|y − x|k P(s, x, t, dy)
|y −x|>ε
Z
= |y − x|2+δ |y − x|k −(2+δ) P(s, x, t, dy)
|y −x|>ε
Z
1
6 2+δ−k |y − x|2+δ P(s, x, t, dy)
ε |y −x|>ε
Z
1
6 2+δ−k |y − x|2+δ P(s, x, t, dy).
ε R d
This implies that assumption (30) is sufficient for the sample paths to
be continuous (k = 0) and for the replacement of the truncated
integrals in (28) and (29) by integrals over Rd (k = 1 and k = 2,
respectively). The definitions of the drift and diffusion coefficients
become:
Xt − Xs
lim E Xs = x = a(x, s) (31)
t→s t −s
and
(Xt − Xs ) ⊗ (Xt − Xs )
lim E Xs = x = b(x, s) (32)
t→s t −s
Now it becomes clear that this condition implies that the probability of
large changes in Xt over short time intervals is small. Notice, on the
other hand, that the above condition implies that the sample paths of a
diffusion process are not differentiable: if they where, then the right
hand side of the above equation would have to be 0 when t − s ≪ 1.
The sample paths of a diffusion process have the regularity of
Brownian paths. A Markovian process cannot be differentiable: we
can define the derivative of a sample paths only with processes for
which the past and future are not statistically independent when
conditioned on the present.
and
Es,x (Xt − Xs ) ⊗ (Xt − Xs ) = b(x, s)(t − s) + o(t − s).
Wt − Ws ∼ N (0, (t − s)I),
Assume furthermore that the functions a(x, s), b(x, s) are continuous
in both x and s. Then u(x, s) ∈ C 2,1 (R × R+ ) and it solves the final
value problem
∂u ∂u 1 ∂2u
− = a(x, s) + b(x, s) 2 , lim u(s, x) = f (x). (33)
∂s ∂x 2 ∂x s→t
Now the final condition follows from the fact that f (x) ∈ Cb (R) and the
arbitrariness of ε.
u(x, s) − u(x, s + h)
=
Z h
1
P(dy, s + h|x, s)u(y, s + h) − u(x, s + h)
h
Z R
1
= P(dy, s + h|x, s)(u(y, s + h) − u(x, s + h))
h R
Z
1
= P(dy, s + h|x, s)(u(y, s + h) − u(x, s)) + o(1)
h |x−y |<ε
Z
∂u 1
= (x, s + h) (y − x)P(dy, s + h|x, s)
∂x h |x−y |<ε
Z
1 ∂2u 1
+ (x, s + h) (y − x)2 P(dy, s + h|x, s)(1 + αε ) + o(1)
2 ∂x 2 h |x−y |<ε
∂u 1 ∂2u
= a(x, s) (x, s + h) + b(x, s) 2 (x, s + h)(1 + αε ) + o(1).
∂x 2 ∂x
Equation (33) follows by taking the limits ε → 0, h → 0.
Pavliotis (IC) StochProc January 16, 2011 127 / 367
Assume now that the transition function has a density p(y, t|x, s). In
this case the formula for u(x, s) becomes
Z
u(x, s) = f (y)p(y, t|x, s) dy.
R
Substituting this in the backward Kolmogorov equation we obtain
Z
∂p(y, t|x, s)
f (y) + As,x p(y, t|x, s) = 0 (37)
R ∂s
where
∂ 1 ∂2
As,x := a(x, s) + b(x, s) 2 .
∂x 2 ∂x
Since (37) is valid for arbitrary functions f (y), we obtain a partial
differential equations for the transition probability density:
∂p(y, t|x, s) ∂p(y, t|x, s) 1 ∂ 2 p(y, t|x, s)
− = a(x, s) + b(x, s) . (38)
∂s ∂x 2 ∂x 2
Notice that the variation is with respect to the "backward" variables
x, s. We can also obtain an equation with respect to the "forward"
variables y, t, the Forward Kolmogorov equation .
Pavliotis (IC) StochProc January 16, 2011 128 / 367
Now we derive the forward Kolmogorov equation or Fokker-Planck
equation. We assume that the transition function has a density with
respect to Lebesgue measure.
Z
P(Γ, t|x, s) = p(t, y|x, s) dy.
Γ
Theorem
(Kolmogorov) Assume that conditions (27), (28), (29) are satisfied and
that p(y, t|·, ·), a(y, t), b(y, t) ∈ C 2,1 (R × R+ ). Then the transition
probability density satisfies the equation
∂p ∂ 1 ∂2
=− (a(t, y)p) + (b(t, y)p) , lim p(t, y|x, s) = δ(x − y).
∂t ∂y 2 ∂y 2 t→s
(39)
∂p(y, t) ∂ 1 ∂2
=− (a(y, t)p(y, t)) + (b(y, t)p(t, y)) , (41)
∂t ∂y 2 ∂y 2
The solution of equation (41), provides us with the probability that the
diffusion process Xt , which initially was distributed according to the
probability density ρ0 (x), is equal to y at time t. Alternatively, we can
think of the solution to (39) as the Green’s function for the PDE (41).
Pavliotis (IC) StochProc January 16, 2011 133 / 367
Quite often we need to calculate joint probability densities. For,
example the probability that Xt1 = x1 and Xt2 = x2 . From the properties
of conditional expectation we have that
p(x1 , t1 , x2 , t2 ) = PXt1 = x1 , Xt2 = x2 )
= P(Xt1 = x1 |Xt2 = x2 )P(Xt2 = x2 )
= p(x1 , t1 |x2 t2 )p(x2 , t2 ).
Using the joint probability density we can calculate the statistics of a
function of the diffusion process Xt at times t and s:
Z Z
E(f (Xt , Xs )) = f (y, x)p(y, t|x, s)p(x, s) dxdy. (43)
In particular,
Z Z
E(Xt X0 ) = yxp(y, t|x, 0)p(x, 0) dxdy.
and
Z
1
lim (y − x) ⊗ (y − x)P(dy, t|x, s) = b(x, s).
t→s t − s |y −x|<ε
and
(Xt − Xs ) ⊗ (Xt − Xs )
lim E X
s = x = b(x, s) (45)
t→s t −s
Notice that from the above definition it follows that the diffusion matrix
is symmetric and nonnegative definite.
3 limt→0 u(t, x ) = f (x ).
n · J(x , t) = 0, on ∂Ω.
p(x , t) = 0, on ∂Ω.
∂p ∂(xp) ∂2p
=α + D 2.
∂t ∂x ∂x
This is the Fokker-Planck equation for the Ornstein-Uhlenbeck
process (Ornstein-Uhlenbeck, 1930). Its solution is
r !
α α(x − e−α(t−s) y)2
pOU (x, t|y, s) = exp − .
2πD(1 − e−2α(t−s) ) 2D(1 − e−2α(t−s) )
d
M1 = −αM1 .
dt
Consequently, the first moment converges exponentially fast to
0:
M1 (t) = e−αt M1 (0).
Or, equivalently:
d
Mn = −αnMn + Dn(n − 1)Mn−2 , n > 2.
dt
This is a first order linear inhomogeneous differential equation.
We can solve it using the variation of constants formula:
Z t
Mn (t) = e−αnt Mn (0) + Dn(n − 1) e−αn(t−s) Mn−2 (s) ds.
0
We have that
lim Mn (t) = hx n iOU .
t→∞
∂ σ2 x 2 ∂ 2
L = µx + .
∂x 2 ∂x 2
Notice that this operator is not uniformly elliptic.
The Fokker-Planck equation of the geometric Brownian motion is:
∂p ∂ ∂ 2 σ2 x 2
=− (µx) + p .
∂t ∂x ∂x 2 2
d σ2
Mn = µn + n(n − 1) Mn , n > 2.
dt 2
The solution of this equation is
σ2
Mn (t) = e(µ+(n−1) 2
)nt
Mn (0), n>2
and
M1 (t) = eµt M1 (0).
Notice that the nth moment might diverge as t → ∞, depending
on the values of µ and σ:
σ2 µ
2 < − n−1 ,
0,
2 µ
lim Mn (t) = Mn (0), σ2 = − n−1 ,
t→∞
σ2 µ
+∞, 2 > − n−1 .
∂p
= ∇ · (∇Vp) + D∆p. (51)
∂t
It is not possible to calculate the time dependent solution of this
equation for an arbitrary potential. We can, however, always
calculate the stationary solution.
Then the Markov process with generator (50) is ergodic. The unique
invariant distribution is the Gibbs distribution
1 −V (x)/D
p(x) = e , (53)
Z
where the normalization factor Z is the partition function
Z
Z = e−V (x)/D dx.
Rd
Theorem
Let p(x, t) be the solution of the Fokker-Planck equation (51), assume
that (52) holds and let ρ(x) be the Gibbs distribution (187). Define
h(x, t) through
p(x, t) = h(x, t)ρ(x).
Then the function h satisfies the backward Kolmogorov equation:
∂h
= −∇V · ∇h + D∆h, h(x, 0) = p(x, 0)ρ−1 (x). (54)
∂t
∇p = ρ∇h − ρhD −1 ∇V
and
∂h
ρ = ρ − ∇V · ∇h + D∆h ,
∂t
from which the claim follows.
L = −∇V (x) · ∇ + D∆
= −D ∇f · ∇hρ dx,
Rd
0 = −Dk∇f k2ρ ,
Theorem
Assume that the potential V satisfies the convexity condition
D 2 V > λI.
Theorem
Let p denote the solution of the Fokker–Planck equation (51) where
the potential is smooth and uniformly convex. Assume that the the
initial conditions satisfy
−Lfn = λn fn , n = 0, 1, . . .
ḣn = −λn hn , n = 0, 1, . . .
Notice that
Z Z
1 = p(x, 0) dx = p(x, t) dx
d Rd
ZR
= h(x, t)Z −1 e−βV dx = (h, 1)ρ = (φ, 1)ρ
Rd
= φ0 .
This expansion, together with the fact that all eigenvalues are
positive (n > 1), shows that the solution of the backward
Kolmogorov equation converges to 1 exponentially fast.
The solution of the Fokker–Planck equation is
∞
!
X
p(x, t) = Z −1 e−βV (x) 1+ e−λn t φn fn .
n=1
∂p
= D∇ · e−V /D ∇ eV /D p . (62)
∂t
Define now ψ(x, t) = eV /2D p(x, t). Then ψ solves the PDE
∂ψ |∇V |2 ∆V
= D∆ψ − U(x)ψ, U(x) := − . (63)
∂t 4D 2
Let H := −D∆ + U. Then L∗ and H have the same eigenvalues. The
nth eigenfunction φn of L∗ and the nth eigenfunction ψn of H are
associated through the transformation
V (x)
ψn (x) = φn (x) exp .
2D
Pavliotis (IC) StochProc January 16, 2011 177 / 367
Remarks
2 The operator that appears on the right hand side of eqn. (63) has
the form of a Schrödinger operator:
−H = −D∆ + U(x).
We calculate
D∇ · e−V /D ∇ eV /D f = D∇ · e−V /D D −1 ∇Vf + ∇f eV /D
= ∇ · (∇Vf + D∇f ) = L∗ f .
−L∗ φn = λn φn .
1
Set φn = ψn exp − 2D V . We calculate −L∗ φn :
From this we conclude that e−V /2D Hψn = λn ψn e−V /2D from which the
equivalence between the two eigenvalue problems follows.
∇U ∇U
H = DA∗ A, A=∇+ , A∗ = −∇ + .
2D 2D
2 These are creation and annihilation operators. They can also be
written in the form
A· = e−U/2D ∇ eU/2D · , A∗ · = eU/2D ∇ e−U/2D ·
where φBn and φFn denote the eigenfunctions of the backward and
forward operators, respectively.
Pavliotis (IC) StochProc January 16, 2011 181 / 367
The OU Process and Hermite Polynomials
The generator of the OU process is
d d2
L = −y +D 2 (64)
dy dy
The OU process is an ergodic Markov process whose unique
invariant measure is absolutely continuous with respect to the
Lebesgue measure on R with density ρ(y) ∈ C ∞ (R)
1 y2
ρ(y) = √ e− 2D .
2πD
Let L2ρ denote the closure of C∞ functions with respect to the norm
Z
kf k2ρ = f (y)2 ρ(y) dy.
R
The space L2ρ is a Hilbert space with inner product
Z
(f , g)ρ = f (y)g(y)ρ(y) dy.
R
Pavliotis (IC) StochProc January 16, 2011 182 / 367
Consider the eigenvalue problem for L:
−Lfn = λn fn .
λn = n, n ∈ N.
H0 (y) = 1,
H1 (y) = y,
H2 (y) = y 2 − 1,
H3 (y) = y 3 − 3y,
H4 (y) = y 4 − 3y 2 + 3,
H5 (y) = y 5 − 10y 3 + 15y.
Hn is a polynomial of degree n.
Only odd (even) powers appear in Hn when n is odd (even).
(fn , fm )ρ = δnm .
and r
D dφ
A− φ(y) = (y).
n dy
Then
A+ fn = fn+1 and A− fn = fn−1 . (66)
Pavliotis (IC) StochProc January 16, 2011 185 / 367
4. The eigenfunctions fn satisfy the following recurrence relation
√ p
yfn (y) = Dnfn−1 (y) + D(n + 1)fn+1 (y), n > 1. (67)
5. The function
λ2
H(y; λ) = eλy − 2
Then the Markov process with generator (140) is ergodic. The unique
invariant distribution is the Maxwell-Boltzmann distribution
1 −βH(p,q)
ρβ (p, q) = e (70)
Z
where
1
H(p, q) = kpk2 + V (q)
2
is the Hamiltonian, β = (kB T )−1 is the inverse temperature and the
normalization factor Z is the partition function
Z
Z = e−βH(p,q) dpdq.
R2d
where {Xj }dj=0 are vector fields (first order differential operators).
Define the Lie algebras
A0 = Lie(X1 , X1 , . . . Xd ),
A1 = Lie([X0 , X1 ], [X0 , X2 ], . . . [X0 , Xd ]),
... = ...,
Ak = Lie([X0 , X ]; X ∈ Ak −1 ), k >1
Define
H = Lie(A0 , A1 , . . . )
Hörmander’s Hypothesis. H is full in the sense that, for every x ∈ Rd
the vectors {H(x) : H ∈ H} span Tx Rd :
span{H(x); H ∈ H} = Tx Rd .
Pavliotis (IC) StochProc January 16, 2011 191 / 367
Theorem (Hörmander/Kolmogorov.)
Under the above hypothesis, the diffusion process Xt with generator L
has a transition density
Z
(Pt f )(x) = p(t, x, y)f (y) dy,
Rd
∂
p(·, x, ·) = L∗y p(·, x, ·), x ∈ Rd (71)
∂t
and
∂
p(·, ·, y) = Lx p(·, ·, y), y ∈ Rd . (72)
∂t
For every x ∈ Rd the function p(·, x, ·) is the fundamental solution
d
R function) of (71), such that, for each f ∈ C0 (R ),
(Green’s
limt→0 Rd p(t, x, y)f (y) dy = f (x).
Pavliotis (IC) StochProc January 16, 2011 192 / 367
Example
Consider the SDE √
ẍ = 2Ẇ .
Write it as a first order system
√
ẋ = y, ẏ = 2Ẇ .
L = X0 + X12 , X0 = y∂x , X1 = ∂y .
which spans Tx R2 .
∂h
= −Ah + γSh
∂t
where
A = p · ∇q − ∇q V · ∇p , S = −p · ∇p + β −1 ∆p .
The operator A is antisymmetric in L2ρ := L2 (R2d ; ρβ (q, p)), whereas S
is symmetric.
∂
Xi∗ = −βpi + .
∂pi
We have that
d
X
−1
S=β Xi∗ Xi .
i=1
where
∂h
= Lh. (74)
∂t
Furthermore
L1 h = −∂q V ∂p h + p∂q h.
We calculate each term on the right hand side of the above
equation separately. For this we will need the formulas
√ √ √
∂p fn = nfn−1 and pfn = nfn−1 + n + 1fn+1 .
with h−1 ≡ 0.
Furthermore
∞
X ∞
X √
∂q V ∂p h = ∂q Vhn ∂p fn = ∂q Vhn nfn−1
n=0 n=0
∞
X √
= ∂q Vhn+1 n + 1fn .
n=0
Lh = L1 + γL1 h
∞
X √
= − γnhn + n + 1∂q hn+1
n=0
√ √
+ n∂q hn−1 + n + 1∂q Vhn+1 fn
or p
q̇ = p, ṗ = −ω02 q − γp + 2γβ −1 Ẇ . (78)
This is a linear equation that can be solved explicitly (The solution
is a Gaussian stochastic process).
βω0 − β p2 − βω02 q 2
ρβ (q, p) dqdp = e 2 2 . (81)
2π
and
We have that
a+ a− = −β −1 ∂p2 + p∂p
and
b + b − = −β −1 ∂q2 + q∂q
Consequently, the operator
Lb = −a+ a− − b + b − (84)
[b + , b − ] = −1, (85b)
[a± , b ± ] = 0. (85c)
L = −Cc + c − − Dd + d − (87)
[c + , c − ] = −1, (88a)
[d + , d − ] = −1, (88b)
[c ± , d ± ] = 0. (88c)
Pavliotis (IC) StochProc January 16, 2011 209 / 367
The operators c ± and d ± should be given as linear combinations
of the old operators a± and b ± . They should be of the form
with q
γ±δ
λ1,2 = , δ= γ 2 − 4ω02 . (90)
2
The eigenvalues satisfy the relations
λ1 + λ2 = γ, λ1 − λ2 = δ, λ1 λ2 = ω02 . (91)
1 p p 1 p p
c+ = √ λ1 a + + λ2 b + , c− = √ λ1 a − − λ2 b − ,
δ δ
(92a)
+ 1 p +
p
+
− 1 p −
p
−
d =√ λ2 a + λ1 b , d = √ − λ2 a + λ1 b .
δ δ
(92b)
± ±
Then c , d satisfy the canonical commutation relations (88) as well
as
[L, c ± ] = −λ1 c ± , [L, d ± ] = −λ2 d ± . (93)
Furthermore, the operator L can be written in the form
L = −λ1 c + c − − λ2 d + d − . (94)
[L, c + ] = −λ1 c + c − c + + λ1 c + c + c −
= −λ1 c + (1 + c + c − ) + λ1 c + c + c −
= −λ1 c + (1 + c + c − ) + λ1 c + c + c −
= −λ1 c + ,
L = −λ1 c + c − − λ2 d + d −
√
λ22 − λ21 + − + − λ1 λ2
= − a a + 0b b + (λ1 − λ2 )a+ b −
δ δ
1p
+ λ1 λ2 (−λ1 + λ2 )b + a−
δ
= −γa+ a− − ω0 (b + a− − a+ b − ),
1 1
λnm = λ1 n + λ2 m = γ(n + m) + δ(n − m), n, m = 0, 1, . . . (95)
2 2
and
1
φnm (q, p) = √ (c + )n (d + )m 1, n, m = 0, 1, . . . (96)
n!m!
[L, (c + )2 ] = L(c + )2 − (c + )2 L
= (c + L − λ1 c + )c + − c + (Lc + + λ1 c + )
= −2λ1 (c + )2
√ n X
X m
− n+m n/2 m/2 1
φnm = n!m!δ 2 λ1 λ2 ×
k!(m − k)!ℓ!(n − ℓ)!
ℓ=0 k =0
k −ℓ
λ1 2
(a+ )n+m−k −ℓ (b + )ℓ+k 1.
λ2
φ00 = 1.
√ `√ √ ´
β λ1 p + λ2 ω0 q
φ10 = √ .
δ
√ `√ √ ´
β λ2 p + λ1 ω0 q
φ01 = √
δ
√ √ √ √ √ √
−2 λ1 λ2 + λ1 β p 2 λ2 + β pλ1 ω0 q + ω0 β qλ2 p + λ2 ω0 2 β q 2 λ1
φ11 = .
δ
√ √
−λ1 + β p 2 λ1 + 2 λ2 β p λ1 ω0 q − λ2 + ω0 2 β q 2 λ2
φ20 = √ .
2δ
√ √
−λ2 + β p 2 λ2 + 2 λ2 β p λ1 ω0 q − λ1 + ω0 2 β q 2 λ1
φ02 = √ .
2δ
Im (λnm) 1
−1
−2
−3
0 0.5 1 1.5 2 2.5 3
Re (λ )
nm
Lb := −A + γS
= −ω0 (b + a− − b − a+ ) + γa+ a−
= −λ1 (c − )∗ (c + )∗ − λ2 (d − ) ∗ (d + )∗,
where
1 p p 1 p p
(c + )∗ = √ λ1 a− + λ2 b − , (c − )∗ = √ λ1 a + − λ2 b +
δ δ
(100a)
1 p p 1 p p
(d + )∗ = √ λ2 a − + λ1 b − , − ∗
(d ) = √ − λ2 a + λ1 b +
+
δ δ
(100b)
Pavliotis (IC) StochProc January 16, 2011 222 / 367
Exercise
1 Show by direct substitution that Lb can be written in the form
Lb = −λ1 (c − )∗ (c + )∗ − λ2 (d − )∗ (d + )∗ .
1
ψnm = √ ((c − )∗ )n ((d − )∗ )m 1. (101)
n!m!
Proof. We will use formulas (98). Notice that using the third and fourth
of these equations together with the fact that c − 1 = d − 1 = 0 we can
conclude that (for n > ℓ)
(c − )ℓ (c + )n 1 = n(n − 1) . . . (n − ℓ + 1)(c + )n−ℓ . (103)
We have
1
Z Z Z Z
φnm ψℓk ρβ dpdq = √ ((c + ))n ((d + ))m 1((c − )∗ )ℓ ((d − )∗ )k 1ρβ dpdq
n!m!ℓ!k !
n(n − 1) . . . (n − ℓ + 1)m(m − 1) . . . (m − k + 1)
= √ ×
n!m!ℓ!k !
Z Z
((c + ))n−ℓ ((d + ))m−k 1ρβ dpdq
= δnℓ δmk ,
since all eigenfunctions average to 0 with respect to ρβ .
Pavliotis (IC) StochProc January 16, 2011 224 / 367
From the eigenfunctions of Lb we can obtain the eigenfunctions of
the Fokker-Planck operator. Using the formula
L∗ (f ρβ ) = ρLf
b
EI(t) = 0
and
E[I(t)|Fs ] = I(s) ∀ t > s,
where Fs denotes the filtration generated by W (s).
1
K
X −1
Istrat (t) := lim f (tk −1 ) + f (tk ) (W (tk ) − W (tk −1 )) , (110)
K →∞ 2
k =1
Definition
By a solution of (105) we mean a Z-valued stochastic process {z(t)}
on t ∈ [0, T ] with the properties:
1 z(t) is continuous and Ft −adapted, where the filtration is
generated by the Brownian motion W (t);
2 h(z(t)) ∈ L1 ((0, T )), γ(z(t)) ∈ L2 ((0, T ));
3 equation (105) holds for every t ∈ [0, T ] with probability 1.
The solution is called unique if any two solutions xi (t), i = 1, 2 satisfy
E|z0 |2 < ∞.
Then the SDE (105) has a unique solution z(t) ∈ C(R+ ; Z) with
"Z #
T
E |z(t)|2 dt < ∞ ∀ T < ∞.
0
again in law.
Hence, if we scale time to s = ct in (105), then we get the equation
dz 1 1 dW
= h(z) + √ γ(z) , z(0) = z0 .
ds c c ds
dx 1
= h(x) + f (x)y, (114a)
dt ε
r
dy αy 2D dV
=− 2 + , (114b)
dt ε ε2 dt
with V being a standard one-dimensional Brownian motion.
D 1 D
= 4 2 2
→
π ε x +α πα2
and, consequently,
y(t) y(s) 2D
lim E = δ(t − s),
ε→0 ε ε α2
which implies the heuristic
r
y(t) 2D dV
lim = . (115)
ε→0 ε α2 dt
Pavliotis (IC) StochProc January 16, 2011 242 / 367
Another way of seeing this is by solving (114b) for y/ε:
r
y 2D dV ε dy
= − . (116)
ε α2 dt α dt
If we neglect the O(ε) term on the right hand side then we arrive,
again, at the heuristic (115).
Both of these arguments lead us to conjecture the limiting Itô SDE:
r
dX 2D dV
= h(X ) + f (X ) . (117)
dt α dt
In fact, as applied, the heuristic gives the incorrect limit.
Theorem
Assume that the initial conditions for y(t) are stationary and that the
function f is smooth. Then the solution of eqn (114a) converges, in the
limit as ε → 0 to the solution of the Stratonovich SDE (118).
D
Lstrat = h(x)∂x + f (x)∂x (f (x)∂x ) .
α
3 Consequently, the Fokker-Planck operator of the Stratonovich
SDE can be written in divergence form:
D 2
L∗strat · = −∂x (h(x)·) + ∂x f (x)∂x · .
α
1 2
1
L = −αy∂ y + D∂y + f (x)y∂x + h(x)∂x
ε2 ε
1 1
=: L0 + L1 + L2 .
ε2 ε
The "fast" process is an stationary Markov process with invariant
density r
α − αy 2
ρ(y) = e 2D . (119)
2πD
The backward Kolmogorov equation is
∂u ε 1 1
= L0 + L1 + L2 u ε . (120)
∂t ε2 ε
We look for a solution to this equation in the form of a power series
expansion in ε:
u ε (x, y, t) = u0 + εu1 + ε2 u2 + . . .
−L0 u0 = 0,
−L0 u1 = L1 u0 ,
∂u0
−L0 u2 = L1 u1 + L2 u0 − .
∂t
The ergodicity of the fast process implies that the null space of the
generator L0 consists only of constant in y. Hence:
u0 = u(x, t).
−L0 u1 = f (x)y∂x u.
1
u1 (x, y, t) = f (x)∂x uy + ψ1 (x, t).
α
In order for the third equation to have a solution we need to require
that the right hand side is orthogonal to the null space of L∗0 :
Z
∂u0
L1 u 1 + L2 u 0 − ρ(y) dy = 0.
R ∂t
We calculate: Z
∂u0 ∂u
ρ(y) dy = .
R ∂t ∂t
Furthermore:
1
Lv = h · ∇v + Γ : ∇∇v. (124)
2
This operator, equipped with a suitable domain of definition, is the
generator of the Markov process given by (105).
The formal L2 −adjoint operator L∗
1
L∗ v = −∇ · (hv) + ∇ · ∇ · (Γv).
2
dV
dV (X (t)) = (X (t)) (f (X (t)) dt + σ(X (t)) ◦ dW (t)) .
dx
Consider the Stratonovich SDE (121) on Rd (i.e. f ∈ Rd ,
σ : Rn 7→ Rd , W (t) is standard Brownian motion on Rn ). The
corresponding Fokker-Planck equation is:
∂ρ 1
= −∇ · (f ρ) + ∇ · (σ∇ · (σρ))). (126)
∂t 2
We can use Itô’s formula to obtain equations for the moments of the
OU process. The generator is:
L = −αx ∂x + λ∂x2 .
Consequently:
Z t
n n
X (t) = x + −αnX (t)n + λn(n − 1)X (t)n−2 dt
0
√ Z t
+n 2λ X (t)n−1 dW .
0
σ2 x 2 2
L = µx ∂x + ∂ .
2 x
The solution to this equation is
σ2
X (t) = X (0) exp (µ − )t + σW (t) . (128)
2
Consequently:
X (t) σ2
log = µ− t + σW (t)
X (0) 2
dXt
= Xt (c − Xt2 ), X0 = x. (129)
dt
dXt √ dWt
= Xt (c − Xt2 ) + 2σ , X0 = x. (130)
dt dt
This equation defines an ergodic Markov process on R: There
exists a unique invariant distribution:
Z
1 1
ρ(x) = Z −1 e−V (x)/σ , Z = e−V (x)/σ dx, V (x) = cx 2 − x 4 .
R 2 4
dXt √ dWt
= Xt (c − Xt2 ) + 2σXt , X0 = x. (131)
dt dt
Where the stochastic differential is interpreted in the Itô sense.
The generator of this process is
γ > −1 ⇒ c > σ.
∂v
= Lv for (z, t) ∈ Z × (0, ∞),
∂t
v = φ for (z, t) ∈ Z × {0} , (133)
Theorem
Consider equation (106) with z(0) a random variable with density
ρ0 (z). Assume that the law of z(t) has a density
ρ(z, t) ∈ C 2,1 (Z × (0, ∞)). Then ρ satisfies the Fokker-Planck
equation
∂ρ
= L∗ ρ for (z, t) ∈ Z × (0, ∞), (134a)
∂t
ρ = ρ0 for z ∈ Z × {0}. (134b)
q γ (t) = λγ q(t/µγ ).
E|p(t)|2 6 C.
∂ρ
= L∗ ρ
∂t
1 1
= (−p∂q ρ + ∂q V (q)∂p ρ) + 2 ∂p (pρ) + β −1 ∂p2 ρ
ε
ε
1 ∗ 1 ∗
=: L + L ρ. (144)
ε2 0 ε 1
remark
This is "almost" the backward Kolmogorov equation with the
difference that we have −L1 instead of L1 . This is related to the
fact that L0 is a symmetric operator in L2 (R2 ; Z −1 e−βH(p,q) ),
whereas L1 is antisymmetric.
L∗0 ρ = L0 (f ρ0 ) = ∂p (f ρ0 ) + β −1 ∂p2 (f ρ0 )
= ρ0 p∂p f + ρ0 β −1 ∂p2 f + f L∗0 ρ0 + 2β −1 ∂p f ∂p ρ0
= −p∂p f + β −1 ∂p2 f ρ0 = ρ0 L0 f .
Similarly,
L0 f0 = 0, (148)
L0 f1 = L1 f0 , (149)
∂f0
L0 f2 = L1 f1 + (150)
∂t
∂fn
L0 fn+1 = L1 fn + , n = 2, 3 . . . (151)
∂t
The null space of L0 consists of constants in p. Consequently,
from equation (148) we conclude that
f0 = f (q, t).
L1 f0 = p∂q f .
L0 f1 = p∂q f .
−∂q V ∂q ψ1 − β −1 ∂q2 ψ1 = 0
∂ρV
= ∂q (∂q V ρV ) + β −1 ∂q2 ρV .
∂t
The limiting SDE lives on the graph associated with the Hamiltonian
system. The domain of definition of the limiting Markov process is
defined through appropriate boundary conditions (the gluing
conditions) at the interior vertices of the graph.
u γ = u0 + γu1 + γ 2 u2 + . . .
L0 u0 = 0, (158a)
∂u0
L0 u1 = −L1 u1 + , (158b)
∂t
∂u1
L0 u2 = −L1 u1 + . (158c)
∂t
.........
Notice that the operator L0 is the backward Liouville operator of the
Hamiltonian system with Hamiltonian
1 2
H= p + V (q).
2
Pavliotis (IC) StochProc January 16, 2011 304 / 367
We assume that there are no integrals of motion other than the
Hamiltonian. This means that the null space of L0 consists of functions
of the Hamiltonian:
N (L0 ) = functions ofH . (159)
Let us now analyze equations (158). We start with (158a); eqn. (159)
implies that u0 depends on q, p through the Hamiltonian function H:
Now we proceed with (158b). For this we need to find the solvability
condition for equations of the form
L0 u = f (161)
where
p 2 = p 2 (H, q) = 2(H − V (q)).
∂u
hp −1 i = β −1 hp −1 i − hpi ∂H u + hpiβ −1 ∂H2 u,
∂t
or,
∂u
= β −1 − hp −1 i−1 hpi ∂H u + γhp −1 i−1 hpiβ −1 ∂H2 u.
∂t
Thus, we have obtained the limiting backward Kolmogorov equation for
the energy, which is the "slow variable". From this equation we can
read off the limiting SDE for the Hamiltonian:
where
1 4 1 2 1
V (x) = x − x + . (169)
4 2 4
This potential has three local minima, a local maximum at x = 0
and two local minima at x = ±1.
−1
−2
−3
0 50 100 150 200 250 300 350 400 450 500
1
V (±1) = 0, V (0) = .
4
Eb = V (b) − V (a).
dx dx
= V ′ (x) = −(V ′ (x))2 < 0.
dt dt
Hence, depending on the initial condition the particle will converge
either to a or c. The particle cannot escape from either state of
local stability.
On the other hand, at high temperatures the particle does not
"see" the potential barrier: it essentially jumps freely from one
local minimum to another.
We will study this problem and calculate the escape rate using by
calculating the mean first passage time.
−Lτ = 1, x ∈ D, (173a)
τ = 0, x ∈ ∂D, (173b)
where L is the generator of the SDE 173.
∂S
= −f (x, t),
∂t
where f (x, t) is the first passage times distribution.
Now we can solve the problem of the escape from a potential well:
the reflecting boundary is at x = a, the left local minimum of the
potential, and the absorbing boundary is at x = b, the local
maximum.
We can also obtain a formula for the reaction rate for γ = O(1):
q
γ2 2 γ
4 − ωb − 2 ω0
κ= exp (−βEb ) . (180)
ωb 2π
hX (t)2 i = 2Dt.
X p2N
PN2 n 1
H(QN , PN , q, p) = + V (QN ) + + kn (qn − λQN )2(186)
.
2 2mn 2
n=1
pn (0)
qn (t) = qn (0) cos(ωn t) + sin(ωn t)
mn ω n
Z t
+ωn λ sin(ωn (t − s))QN (s) ds.
0
pn (0)
qn (t) = qn (0) cos(ωn t) + sin(ωn t) + λQN (t)
mn ω n
Z t
−λQN (0) cos(ωn t) − λ cos(ωn (t − s))Q̇N (s) ds.
0
f 2 (ωn )
kn = ,
N 2b
where the function f (ωn ) decays sufficiently fast at infinity.
We can rewrite the dissipation and noise terms in the form
N
X
RN (t) = f 2 (ωn ) cos(ωn t) ∆ω
n=1
and
N
X √
FN (t) = f (ωn ) (ξn cos(ωn t) + ηn sin(ωn t)) ∆ω,
n=1
where ∆ω = N a /N.
Pavliotis (IC) StochProc January 16, 2011 348 / 367
Using now properties of Fourier series with random
coefficients/frequencies and of weak convergence of probability
measures we can pass to the limit:
dQ
= P,
dt
dP
= −V ′ (Q) + λZ ,
dt
dZ p dW
= −αZ − λP + 2αβ −1 ,
dt dt
where Z (0) ∼ N (0, β −1 ).
The process {Q(t), P(t), Z (t)} ∈ R3 is Markovian.
It is a degenerate Markov process: noise acts directly only on one
of the 3 degrees of freedom.
We can eliminate the auxiliary process Z by taking an appropriate
distinguished limit.
dQ
= P,
dt √
dP γ
= −V ′ (Q) + Z,
dt ε r
√
dZ 1 γ 2β −1 dW
= − 2Z − P+ .
dt ε ε ε2 dt
We can use tools from singular perturbation theory for Markov
processes to show that, in the limit as ε → 0, we have that
1 p dW
Z → 2γβ −1 − γP.
ε dt
Thus, in this limit we obtain the Markovian Langevin Equation
(R(t) = γδ(t))
p dW
Q̈ = −V ′ (Q) − γ Q̇ + 2γβ −1 . (201)
dt
Pavliotis (IC) StochProc January 16, 2011 354 / 367
Whenever the GLE (195) has "finite memory", we can represent it
as a Markovian SDE by adding a finite number of additional
variables.
These additional variables are solutions of a linear system of
SDEs.
This follows from results in approximation theory.
Consider now the case where the memory kernel is a bounded
analytic function. Its Laplace transform
Z +∞
b
R(s) = e−st R(t) dt
0
can be represented as a continued fraction:
b ∆21
R(s) = , γi > 0, (202)
∆22
s + γ1 + ...
Since R(t) is bounded, we have that
b
lim R(s) = 0.
s→∞
with
β −1 R(t − s) = hF1 (t)F1 (s)i.
Q̇ = P,
Ṗ = −V ′ (Q) + λZ1 (t),
Ż1 = −α1 Z1 + Z2 ,
p
Ż2 = −α2 Z2 − λP + 2β −1 α2 Ẇ2 .
where The function ρ(x) models the coupling between the particle
and the field.
This coupling is influenced by the dipole coupling approximation
from classical electrodynamics.
with appropriate definitions for the memory kernel and the noise,
which are related through the fluctuation-dissipation theorem.
PfN+1 = f , PfN+1 = h.
∂f
i = PL(f + h), (216a)
∂t
∂h
i = (I − P)L(f + h). (216b)
∂t