Professional Documents
Culture Documents
2020 Notes Numprofin
2020 Notes Numprofin
Jean-Francois CHASSAGNEUX∗,
∗
Université de Paris, U.F.R. de Mathématiques, chassagneux@lpsm.paris
1
References
[1] Frédéric Abergel and Rémi Tachet. A nonlinear partial integro-differential equa-
[2] Yacine Ait-Sahalia. Testing continuous-time models of the spot interest rate.
[3] Aurélien Alfonsi et al. Affine diffusions and related processes: simulation, theory
[4] Aurélien Alfonsi, Benjamin Jourdain, and Arturo Kohatsu-Higa. Pathwise opti-
mal transport bounds between a one-dimensional diffusion and its euler scheme.
[5] Vlad Bally and Gilles Pages. Error analysis of the optimal quantization al-
106(1):1–40, 2003.
[6] Vlad Bally, Gilles Pages, et al. A quantization algorithm for solving multidi-
2003.
[7] Vlad Bally and Denis Talay. The law of the euler scheme for stochastic differ-
[8] Denis Belomestny et al. Solving optimal stopping problems via empirical dual
2
[9] Denis Belomestny and John Schoenmakers. Projected particle methods for
Springer, 1978.
[12] François Bolley. Separability and completeness for the wasserstein distance. In
[13] Mireille Bossy and Denis Talay. Convergence rate for the approximation of the
[14] Mireille Bossy and Denis Talay. A stochastic particle method for the mckean-
[16] Bruno Bouchard, Ivar Ekeland, and Nizar Touzi. On the malliavin approach to
3
[17] Bruno Bouchard and Nizar Touzi. Discrete-time approximation and monte-carlo
[18] Gerard Brunick and Steven Shreve. Mimicking an itô process by a solution of a
1628, 2013.
[19] René Carmona, François Delarue, Gilles-Edouard Espinosa, and Nizar Touzi.
[20] René Carmona, François Delarue, et al. Probabilistic Theory of Mean Field
[21] René Carmona, Jean-Pierre Fouque, and Li-Hsien Sun. Mean field games and
[22] Jean-François Chassagneux. Linear multistep schemes for bsdes. SIAM Journal
[24] Jean-François Chassagneux, Lukasz Szpruch, and Alvin Tse. Weak quantitative
4
[25] Emmanuelle Clément, Damien Lamberton, and Philip Protter. An analysis of
a least squares regression method for american option pricing. Finance and
[27] Dan Crisan and Konstantinos Manolarakis. Solving backward stochastic differ-
backward sdes and simulation with the cubature method. The Annals of Applied
[29] Dan Crisan, Konstantinos Manolarakis, and Nizar Touzi. On the monte carlo
[31] Nicole El Karoui, Shige Peng, and Marie Claire Quenez. Backward stochastic
[32] Michael B Giles. Multilevel monte carlo path simulation. Operations Research,
56(3):607–617, 2008.
5
[33] Paul Glasserman. Monte Carlo methods in financial engineering, volume 53.
[35] Emmanuel Gobet. Weak approximation of killed diffusion using euler schemes.
[36] Emmanuel Gobet. Monte-Carlo methods and stochastic processes: from linear
Vázquez. Stratified regression monte-carlo scheme for semilinear pdes and bsdes
38(6):C652–C677, 2016.
[39] Emmanuel Gobet and Plamen Turkedjiev. Linear regression mdp scheme for
6
[41] S Graf, H Luschgy, et al. Asymptotics of the quantization errors for self-similar
[42] Carl Graham, Thomas G Kurtz, Sylvie Méléard, Philip Protter, and Mario
Lectures Given at the 1st Session of the Centro Internazionale Matematico Es-
tivo (CIME) Held in Montecatini Terme, Italy, May 22-30, 1995. Springer,
2006.
[43] J Guyon and P Henry-Labordere. Non linear pricing. crc financial mathematics,
2014.
[44] Julien Guyon and Pierre Henry-Labordere. The smile calibration problem
[45] Julien Guyon and Pierre Henry-Labordère. Being particular about calibration.
cesses having an itô differential. Probability theory and related fields, 71(4):501–
516, 1986.
[47] Martin B Haugh and Leonid Kogan. Pricing american options: a duality ap-
[48] Martin Hutzenthaler, Arnulf Jentzen, and Peter E Kloeden. Strong and weak
7
Society A: Mathematical, Physical and Engineering Sciences, 467(2130):1563–
1576, 2010.
switching local volatility model and new fake brownian motions. arXiv preprint
arXiv:1607.00077, 2016.
[52] Vassili N Kolokoltsov. Nonlinear Markov processes and kinetic equations, vol-
[53] Hiroshi Kunita. Stochastic flows and stochastic differential equations, vol-
[54] Daniel Lacker, Mykhaylo Shkolnikov, and Jiacheng Zhang. Inverting the marko-
[56] Jean-Philippe Lemor, Emmanuel Gobet, Xavier Warin, et al. Rate of con-
8
[57] A Lipton. The vol-smile problem. Risk, 15:61–65, 2002.
[58] Yating Liu. Optimal Quantization: Limit Theorems, Clustering and Simulation
14(1):113–147, 2001.
[60] Harald Luschgy, Gilles Pagès, et al. Functional quantization rate and mean
and their applications. Number 1702. Springer Science & Business Media, 1999.
[64] Gilles Pagès and Abass Sagna. Improved error bounds for quantization based
numerical schemes for bsde and nonlinear filtering. Stochastic Processes and
[65] Etienne Pardoux and Shige Peng. Adapted solution of a backward stochastic
9
[66] Etienne Pardoux and Shige Peng. Backward stochastic differential equations
[67] Étienne Pardoux and Aurel Răşcanu. Stochastic differential equations, Back-
[68] Huyên Pham. Continuous-time stochastic control and optimization with finan-
cial applications, volume 61. Springer Science & Business Media, 2009.
[70] Daniel Revuz and Marc Yor. Continuous martingales and Brownian motion,
[73] Denis Talay and Luciano Tubaro. Expansion of the global error for numeri-
[74] John N Tsitsiklis and Benjamin Van Roy. Regression methods for pricing com-
703, 2001.
10
[75] Cédric Villani. Optimal transport: old and new, volume 338. Springer Science
[76] Jianfeng Zhang et al. A numerical scheme for bsdes. the annals of applied
11
Notations
12
Part I
Handouts
13
1 Introduction
These handouts have non-empty intersection with some textbooks: [51, 36, 63, 33]
14
2 Review of the linear case
2.1.1 SDEs
|b(t, x) − b(t0 , x0 )| + |σ(t, x) − σ(t0 , x0 )| ≤ L |x − x0 | + |t − t0 | .
Theorem 2.1. (strong Existence and uniqueness for SDEs) Under (HL)(i), there
exists a unique1 continuous F-adapted process X taking its values in Rd such that
Z t d Z
X t
Xti = xi0 + i
b (s, Xs ) ds + σ ij (s, Xs ) dWsj (2.2)
0 j=1 0
1
Up to indistinguishability.
15
2.1.2 Useful estimates under (HL)
• For all p ≥ 1,
" #
p
E sup |Xt | ≤ CT (1 + E[|X0 |p ]). (2.3)
t∈[0,T ]
• Time regularity:
" #
2
max E sup |Xt − Xti | ≤ C|ti+1 − ti |. (2.4)
i t∈[ti ,ti+1 ]
Stochastic flow Observe that we can define for all (s, x) ∈ [0, T ]×Rd the solution
to
Z t Z t
Xts,x = x + b(r, Xrs,x )dr + σ(r, Xrs,x )dWr , s ≤ t ≤ T
s s
• Let 2 ≤ p < ∞. There exists a constant Cp such that for all (s, x, t) and
(s0 , x0 , t0 ),
h 0 0
i p p
E |Xts,x − Xts0 ,x |p ≤ Cp |x − x0 |p + (1 + |x|p ∨ |x0 |p )(|s − s0 | 2 + |t − t0 | 2 ) .
(2.6)
16
• The mapping (s, x) 7→ X s,x with values in Sc2 is continuous.
r,x
Xtr,x = Xts,Xs , P − a.s.
17
2.1.3 Link with PDEs
1 2
LX ϕ(t, x) = b(t, x)∂x ϕ(t, x) + Tr ∂xx ϕa (t, x) (2.7)
2
for ϕ ∈ C 2 , where a := σσ † .
that (HL)(i) holds true, g ∈ Cp0 and that there is a u a Cp1,2 ([0, T )×Rd )∩Cp0 ([0, T ]×
then
h i
u(t, x) = E g(XTt,x )
,→ uniqueness result...
for all υ ∈ Rd .
• This assumption allows to prove the smoothness of the function u (for t < T )
18
2.1.4 Financial setting
d
X
dSti = rSti dt + σ ij (St )dWtj
j=1
(investment in risky asset), V p,φ is the portfolio value with initial value p and
• In “perfect market”
p(G),φ∗
p(G) = EQ e−rT G = V0 (2.11)
p(G) = u(0, S0 ) ,
where u is solution (in some sense) to the following parabolic PDE, see e.g.
19
[15]:
∂t u + LS u = ru on [0, T ) × Rd
(2.12)
u(·, T ) = g(·)
20
2.2 Euler Scheme for SDEs
To compute the option price, one can solve numerically the PDE or try to evaluate
the expectation, using e.g. Monte Carlo methods. We will indeed solve the PDE
but using probabilistic methods essentially, meaning that they are inspired by the
SDE.
of the SDE.
Definition 2.1. An Euler approximation of the equation (2.1) associated with the
21
scheme
Xtπi+1 = Xtπi + b ti , Xtπi (ti+1 − ti ) + σ ti , Xtπi Wti+1 − Wti (2.13)
22
Remark 2.1. .
(i) If σ ≡ 0 then (2.13) reduces to the deterministic Euler scheme for ODEs.
(ii) Wti+1 − Wti is a Gaussian random variable with zero mean and variance ti+1 −
Definition 2.2. We denote by {Xtπ , t ∈ [0, T ]}, the continuous Euler Scheme.
Xtπi+1 = Xtπi + b ti , Xtπi (ti+1 − ti ) + σ ti , Xtπi Wti+1 − Wti
Xtπ = Xtπi + b ti , Xtπi (t − ti ) + σ ti , Xtπi (Wt − Wti ) . (2.14)
1 2
L̄(s,z) ϕ(t, x) = b(s, z)∂x ϕ(t, x) + Tr ∂xx ϕ(t, x)a(s, z) (2.15)
2
Moment estimate
23
2.2.2 Weak convergence for vanilla options
1 √
E |f (Xtπ ) − f (Xtπi )|q q ≤ Lf h (2.18)
Li := 2
sup {|∂x u(t, ·)|∞ + |∂xx 3
u(t, ·)|∞ + |∂xxx u(t, ·)|∞ }
t∈[ti ,ti+1 ]
and by L the Lipschitz constant of b and a (to simplify in the proof σ is bounded)
We observe that
n−1
X h i
w = E u(ti+1 , Xtπi+1 ) − u(ti , Xtπi ) .
i=0
h i
E u(ti+1 , Xtπi+1 ) − u(ti , Xtπi ) =
Z ti+1
π π π 1 2 π 2 π
E ∂t u(t, Xt ) + b(Xti )∂x u(t, Xt ) + σ (Xti )∂xx u(t, Xt ) dt
ti 2
h i
E u(ti+1 , Xtπi+1 ) − u(ti , Xtπi ) = (2.19)
Z ti+1
1
E {b(Xtπi ) − b(Xtπ )}∂x u(t, Xtπ ) + {a(Xtπi ) − a(Xtπ )}∂xx
2
u(t, Xtπ ) dt.
ti 2
24
For the first term in the RHS we compute
|E {b(Xtπi ) − b(Xtπ )}∂x u(t, Xtπ ) | =
|E {b(Xtπi ) − b(Xtπ )}{∂x u(t, Xtπ ) − ∂x u(t, Xtπi )} + {b(Xtπi ) − b(Xtπ )}∂x u(t, Xtπi ) |
Z ti
i π π
≤ LL hi + E ∂x u(t, Xti ) L̄z=Xtπ b(Xs )ds
i
t
where we used (2.18) and Cauchy-Schwarz inequality to get the upper bound for the
first term in the RHS of the above inequality. For the second term, we apply Ito’s
Formula:
Z ti Z ti
π π π
π
E b(Xti ) − b(Xt )∂x u(t, Xti ) = E ∂x u(t, Xti ) L̄ z=Xtπ b(Xsπ )ds + ∂x b(Xsπ )σ(Xtπi )dWs
i
t t
Conditioning with respect to Fti cancels the strochastic integral term. Using the
assumptions on b, we get
|E {b(Xtπi ) − b(Xtπ )}∂x u(tti , Xtπi ) | ≤ CL hi .
We thus obtain
|E {b(Xtπi ) − b(Xtπ )}∂x u(t, Xtπ ) | ≤ CL hi .
For the second term in (2.19), we perform similar computations and obtain the same
n−1
X
|w | ≤ C Li h2i , (2.20)
i=0
≤ Ch . (2.21)
25
Extensions .
Theorem 2.4. If u ∈ C ∞ ,
n
X
E[g(XTπ )] = E[g(XT )] + Ci hi + O(hn+1 )
i=1
• For the Euler Scheme with Brownian increments and with condition on the diffu-
26
2.3 Implementation using Monte Carlo Methods
• Assuming we know how to sample from ST , the price p0 = E e−rT g(ST ) < ∞
is approximated by
N
1 X −rT
p̂N
0 = e g(STj ) ,
N
j=1
where the STj are iid random variables with same law as ST .
,→ empirical mean.
Theorem 2.5. (LLN) Since (g(STj ))j is a sequence of integrable and iid random
variables, then
p̂N
0 → p0 a.s.
• We need to assess the accuracy of the previous estimate. We can use the
following L2 -estimate:
Theorem 2.6. Assume that (g(STj ))j a sequence of square integrable and iid random
N V ar(e−rT g(ST ))
Var p̂N
0 = E |p̂0 − p0 |
2
= .
N
27
• Remark: If X1 ∈ L2 , then
N
1 X rT
V̂0N := (e g(STj ) − p̂N
0 )
2
N −1
j=1
is an unbiased estimator of Var erT g(ST ) .
28
• We can also describe the distribution of p̂N
0 , at least asymptotically.
Theorem 2.7. (CLT) Assume that (g(STj ))j is a sequence of square integrable and
q
N V̂0N
iid random variables. We assume that Var[g(ST )] > 0 and set σ̂0 := N then
√ p̂N − p0
N 0 N 1σ̂N >0 → N (0, 1) in distribution.
σ̂0 0
√ p̂N − p0
P( N 0 N < z α2 ) → 1 − α
σ̂0
where for ∈ [0, 1], z denotes the 1 − quantile of the standard normal distribution
σ̂0N N σ̂0N
p0 ∈ [p̂N
0 − z α √
2
, p̂ 0 + z α √
2
]
N N
29
is closed to 1 − α.
val.
• BUT in practice, we do not know how to sample from ST and we use a dis-
• We then compute
N
1 X −rT
p̂π,N
0 := e g(STπ,j )
N
j=1
p̂π,N
0 − p0 = M C + w (2.22)
with M C := p̂π,N
0 − E e−rT g(STπ )
and w = E e−rT g(STπ ) − E e−rT g(ST )
• The study of the MC error can be done as previously as soon as E |g(STπ )|2 <
∞.
• In estimating the quantity p0 = E e−rT g(ST ) , one faces a tradeoff between
30
• To take this into account, one tries to minimise the Mean Square Error.
M SE := E |α − α̂|2
We observe
Indeed,
E |α − α̂|2 = E |α − E[α̂] |2 + 2E[(α − E[α̂])(E[α̂] − α̂)] + E |α̂ − E[α̂] |2
= |α − E[α̂] |2 + V ar(α̂) .
1
M SE ∼c h2 + .
N
,→ overall C ∼c N/h
• The goal is to minimise the MSE taking into account the computational cost:
c2 c3 N
min(c1 h2 + ) s.t. =C.
h,N N h
31
1
• This leads to, setting h2 ∼c N,
√ 1
M SE = O(C − 3 ) .
√
• In other words, to reach a precision M SE = O(), one has
C = O(−3 ) .
• Note: To give (asymptotic) confidence interval for p0 , one needs to make the
discretisation error negligeable with respect to the stat error and to chose then
32
2.4 Implementation using quantisation of Brownian increments
h i h i
ci )k = E (∆Wi )k , .
E (∆W (2.24)
Proposition 2.2. Assume that g and u(t, ·) is Cb4 (with bounds uniform in time) and
h i
c has the matching moment property up to order M = 3 and E (∆W
that ∆W ci )4 =
O(|π|2 ) then
h i
bTπ ) − E[g(XT )] | ≤ C|π| .
ˆw := |E g(X (2.25)
p
ci = ± hi ) = 1 .
P(∆W
2
h i h i h i
ci )2 = E (∆Wi )2 = hi and E (∆W
E (∆W ci )2k+1 = E (∆Wi )2k+1 = 0 , (2.26)
Proof. 1. We drop the π in the proof and denote the Euler scheme with increment
t ,X
i
bt
i bt + ∆X̄i with ∆X̄i := b̂i hi + σ̂i ∆Wi .
X̄ti+1 =X i
33
We observe that
n−1
X h i
ˆw = bt ) − u(ti , X
E u(ti+1 , X bt )
i+1 i
i=0
n−1
X
b ti ,X
bt ti ,X
bt
b
= E u(ti+1 , Xti+1 ) − u(ti+1 , X̄ti+1 ) + E u(ti+1 , X̄ti+1 ) − u(ti , Xti )
i i
i=0
The second term has already been studied in the proof of Theorem 2.3. We just have
to study the first one. We proceed by doing an expansion of the smooth function
bt + λ∆X
Introducing λ 7→ ϕ(X bi ), and performing a Taylor expansion, we compute
i
3
X b i )k Z 1 3
b t + ∆X
ϕ(X bi ) = ϕ(k) b
(Xti )
(∆X
+ bt + λ∆X
ϕ(4) (X bi )4 (1 − λ) dλ ,
bi )(∆X
i i
k! 0 6
k=0
and similarly
3
X Z 1
t ,X bt ) (∆X̄i )k bt + λ∆X̄i )(∆X̄i )4 (1 − λ)3
ϕ(k) (X ϕ(4) (X
i
bt
i
ϕ(X̄ti+1 )= i + i dλ .
k! 0 6
k=0
h i
b ti ,X
E ϕ(Xti+1 ) = E ϕ(X̄ti+1 ) + O(h2i ) .
bt
i
The proof is then concluded observing that, from the previous expansion,
b ti ,X
E u(ti+1 , Xti+1 ) − u(ti+1 , X̄ti+1 ) = O(h2i )
bt
i
Tree methods
• When using discrete random variables for the increment, it is possible to cal-
34
• There is no statistical error but computation becomes rapidly untractable (un-
h i
ūi (x) = E ūi+1 (X̂tti+1
i ,x
)
1 √ 1 √
= ūi+1 (x + σ h) + ūi+1 (x − σ h)
2 2
where ui (·) (resp. ui+1 (·)) stands for the approximation of uti (·) (resp. uti+1 (·))
35
Figure 2: MC simulation for a put (gauss inc)
36
Figure 3: MC simulation for a digital (gauss inc)
37
Figure 5: Bias for a digital (disc inc) - no variance
JUMP
38
2.5 Strong convergence
Proposition 2.3. Assume that b and σ are Lipschitz continuous in t and x, then
" #
E sup |Xt − Xtπ |2 ≤ C|π| .
t∈[0,T ]
Z u Z u
δXu = b(Xs̄ ) − b(Xs̄π ) ds + σ(Xs̄ ) − σ(Xs̄π ) dWs + T (u) (2.27)
0 0
Z u Z u
T (u) = b(Xs ) − b(Xs̄ ) ds + σ(Xs ) − σ(Xs̄ ) dWs . (2.28)
0 0
We then compute
Z t Z u
2
E sup |δXu | ≤ C(E |b(Xs̄ ) − b(Xs̄π )|2 ds + sup | σ(Xs̄ ) − σ(Xs̄π ) 2 2
dWs | + sup |T (u)| .
u≤t 0 u≤t 0 u≤t
Applying BDG inequality for the stoc. int. term and then using the Lipschitz
Z t
2 2 2
E sup |δXu | ≤ C(E |δXs̄ | ds + sup |T (u)| ) , (2.29)
u≤t 0 u≤t
leading to
"Z #
t
E sup |δXu |2 ≤ C(E sup |δXu |2 ds + sup |T (u)|2 ) . (2.30)
u≤t 0 u≤s u≤T
39
This step shows that the global error is controlled by the (global) truncation error!
Z T Z u
2 2
sup |T (u)| ≤ C( b(Xs ) − b(Xs̄ ) ds + sup | σ(Xs ) − σ(Xs̄ ) dWs |2 ) .
u≤T 0 u≤T 0
(2.32)
• The result above yields a naturally a control on the strong approximation error at
• Let assume a uniform grid π with h = T /n. We now report the strong error
Remark 2.2. Obviously, the same control holds true with Xti instead of Xtπi . In
fact it is sharp for both, taking the specific case of X = W , see [63], Chapter 7.
40
2.5.2 Non globally Lipschitz case
• Existence and uniqueness results are more complicated to obtain when the coeffi-
• We consider below SDEs strong uniqueness & existence result: see the book by
Khasminskii [50] for a Lyapounov function approach or the book by Alfonsi [3] for
Explicit Euler scheme diverges Explicit Euler Scheme for SDEs with super-
Existence and uniqueness comes from Theorem 2.4.1 in [50]. It is also the case that
• Euler scheme is, as usual, on a uniform time grid π with n time steps, h = T /n,
Proof. The proof is based on identifying an event with exponential small probability
41
where rn = max( h3 , 2).
i−1
1. By induction we show that, |Xtπi | ≥ rn2 on Ωn .
Now assume the property holds true for i < n. Observing that |a − (b + c)| ≥
|a| − |b + c| ≥ |a| − |b| − |c|, we get, using the scheme definition (2.35),
≥ h|Xtπi |3 − 2|Xtπi |2 ,
where we used the fact that |Xtπi | ≥ 1 for the last inequality. We then compute
i
|Xtπi+1 | ≥ |Xtπi |2 (hrn − 2) ≥ (rn )2 .
2. We now compute
1
Observing that supt∈[0,T ] |Wt | ≤ 2 implies that |Wt − Ws | ≤ 1 for any (s, t) ∈ [0, T ]2 ,
we get
1
P(Ωn ) ≥ P( sup |Wt | ≤ )P(|Wh | ≥ rn )
t∈[0,T ] 2
rn
≥ cP(|W1 | ≥ √ ) for c > 0
h
2
xe−x
P(|W1 | ≥ x) ≥ (2.37)
4
(rn )2
P(Ωn ) ≥ ce− h
42
3. Combining step 1. with the above inequality, we conclude
n−1 n3
E[XTπ ] ≥ c22 e9 T 3 → +∞ when n → ∞
moreover E[|XT |] < ∞ so that the result comes from a direct application of (2.36).
• The proof above can be extended to more general situation see Theorem 1 in [48].
Namely, let
Euler scheme.
Assume moreover, P(σ(ξ) 6= 0) > 0 and that for some constant C ≥ 1, β > α > 1,
|x| ≥ C,
|x|β
max(|µ(x)|, |σ(x)|) ≥ and min(|µ(x)|, |σ(x)|) ≤ C|x|α
C
then
lim E[|XTπ |] = +∞ .
n→∞
43
• Depending on the SDEs various solutions can be proposed, among them the use
• Typical example of symptomatic financial models are short rate models, as Aït-
a−1
dXt = − a0 + a1 Xt − a2 Xt dt + γXtρ dWt and X0 > 0 ,
%
Xt
ai ≥ 0, ρ, % > 1.
44
2.6 Path-dependent options
like to evaluate
F ρ of F .
3. Barrier option
45
for some positive constant L.
t − ti
Xt`,π = Xtπi + (X π − Xtπi )
ti+1 − ti ti+1
• The continuous version of the Euler scheme X π can be used if F depends only on
quantities of the path that can be simulated (see e.g. next section). Then, we have,
1
|E[F (X· ) − F (X·π )] | ≤ CL √
n
• The lower bound in this case is given by the weak error analysis and is of order
2
1
n. The best known upper bound [4] is C n− 3 + in dimension one under an uniform
ellipticity assumption..
Example: Weak convergence for barrier options (1d case) See [35] (or 2019
lecture notes)
46
2.7 Acceleration methods
• When there is no bias: one should implement Variance reduction methods (see
e.g. )
2α+1
C = O(− α ),
√
for a M SE = O().
47
2.7.2 Multi-Level Monte Carlo
Introduced in [32].
We want to approximate
T
h` = , ` = 0, 1, . . . , L .
2`
• We denote by
`
G` = g(XTπ )
• We observe
L
X
E[GL ] = E[G0 ] + E[G` − G`−1 ]
`=1
RHS in a way which minimises the overall variance for a given computational
cost.
• Y0 is an estimator for E[G0 ] using N0 sample paths, and Y l for E[G` − G`−1 ]
48
PL
• We observe, setting Y = Y 0 + `=1 Y
` (main estimator)
h i V`
L
X
Var Y ` = and Var[Y ] = N`−1 V`
N`
`=0
L
X
N` h−1
` .
`=0
1
constants s.t. α ≥ 2 min(γ, β), that
then there exists c4 > 0, such that there are values L and N` s.t. M SE :=
E |Y − E[G] |2 < 2 for < 1/e with
c4 −2 , β>γ
C ≤ c4 −2 (log(1/))2 , β=γ
c4 −2−(γ−β)/α , 0<β<γ
1
2 in the strong sense, we have
49
we obtain
E[GL − G] = O()
Var[Y ] = O(2 )
M SE = O(2 )
50
Proof. We first observe that
L
!
X h`
M SE ≤ C + h2L (2.38)
N`
`=0
L
X N`
=C.
h`
`=0
We minimise the RHS of (2.38) with L being fixed for now. Observing that the
problem is similar to
L
X L
X
1
min s.t. x` = 1
x`
`=0 `=0
1
whose solution is simply x` = L+1 , we compute that
h` C (L + 1)2
N` = and M SE = + h2L . (2.39)
L+1 C
log(1/)
L= and C ∼c −2 (log(1/))2 . (2.40)
log(2)
51
Figure 6: MLMC for European option (M=4), see [32]
0 0
−2
−2
−4
log variance
log |mean|
−4
−6
M
M
−6
−8
P
l
−8 P P− P
l −10 l l−1
P− P Y −Y /M
l l−1 l l−1
−10 −12
0 1 2 3 4 0 1 2 3 4
l l
10 1
10 10
ε=0.00005
ε=0.0001
8 ε=0.0002
10 10
0
ε=0.0005
ε=0.001
ε2 Cost
Nl
6
10 −1
10
Std MC
4 Std MC ext
10
−2 MLMC
10 MLMC ext
0 1 2 3 4 −4 −3
10 10
l ε
16
2.8 High order schemes
scheme, one has to add more term in the expansion of the coefficient.
define X π by
• X0π = X0
• For 0 ≤ i ≤ n − 1,
1
Xtπi+1 = Xtπi + b(Xtπi )hi + σ(Xtπi )∆Ŵi + σ(Xtπi )σ 0 (Xtπi ) (∆Ŵi )2 − hi
2
1 1 00
+ b0 (Xtπi )σ(Xtπi ) + b(Xtπi )σ 0 (Xtπi ) + σ (Xtπi )σ 2 (Xtπi ) ∆Ŵi hi
2 2
1 0 π 1
bb (Xti ) + b00 σ 2 (Xtπi ) h2i
2 2
p 1 2
P(∆Ŵi = ± 3hi ) = , P(∆Ŵi = 0) = .
6 3
53
3 Computing sensitivities in the linear case
• We focus on the Delta but this approach can be applied to other "greeks"
h i
• The price of the european option is given by u(t, x) = E g(XTt,x ) where X is
solution to an SDE of type (2.1) (with b = 0, d = 1), its Delta is ∂x u(t, x).
u(0, x + ) − u(0, x)
∂x u(0, x) ≈ ∆F () :=
Remark: one can consider also backward difference. Same properties here.
u(0, x + ) − u(0, x − )
∂x u(0, x) ≈ ∆C () :=
2
• Estimator (for the central difference): let Xix− , Xix+ be iid samples of XT0,x−
N
1 X
ˆ N () :=
∆ C g(Xix+ ) − g(Xix− )
2N
i=1
3.1.1 Bias
54
Proof. For the central difference: If u(0, .) is C 3 , we simply observe
1 2
u(0, x + ) = u(0, x) + ∂x u(0, x) + ∂xx u(0, x)2 + O(3 )
2
1 2
u(0, x − ) = u(0, x) − ∂x u(0, x) + ∂xx u(0, x)2 + O(3 )
2
55
3.1.2 Variance
ˆ N.
We focus on ∆ C
(i) Using independent sets of random numbers for Xix− and Xix+ :
0,x+
ˆN 1 V ar(g(XT )) + V ar(g(XT0,x− ))
V ar(∆ C ()) =
N 2 4
ˆ N ()) = 1
V ar(∆ C (V ar(p̂N (x + )) + V ar(p̂N (x − )))
42
(ii) Using the same set of random numbers for Xix− and Xix+ and assuming
g, b and σ are C 1 :
ˆN 1 g(XTx+ ) − g(XTx− )
V ar(∆C ()) = V ar
N
56
3.2 Tangent process approach
Theorem 3.1. If b and σ are Cb2 2 then for all t ∈ [0, T ], the mapping x 7→ Xtx is
Z t d Z
X t
∇Xtx = Id + ∇b(Xsx )∇Xsx ds + ∇σ .j (Xsx )∇Xsx dWsj .
0 j=1 0
Xtx+ − Xtx
lim = ∇Xtx ,
→0
and x 7→ ∇X x is continuous.
(d = 1)
To obtain the a.s. properties claimed in the statement of the Theorem, one has to
use the Kolmogorov continuity criterion, see e.g. Chapter 1, Theorem 1.8 in [70].
X x+ −X x
1. We define ∆ X := X x+ − X x , δ X := and compute
Z t Z t
∆ Xt = + b̃u ∆ Xu dWu + σ̃u ∆ Xu dWu ,
0 0
2
b, σ Cb1 with α-Holder continuous derivatives for α > 0 is sufficient, see [53].
57
R1 R1
where we set σ̃u = 0 σ 0 (Xux+ + λ∆ Xu )dλ and b̃u = 0 b0 (Xux+ + λ∆ Xu )dλ.
Since σ and b have bounded derivatives, we observe that σ̃ and b̃ are bounded. We
compute, for p ≥ 2
Z s Z s
p p−1 p
E sup |∆ Xs | ≤ 3 ( + E sup | b̃u ∆ Xu du|p + E sup | σ̃u ∆ Xu dWu |p )
s≤t s≤t 0 s≤t 0
Z t Z s
p p
≤ C( + E sup |∆ Xs | du + E sup | σ̃u ∆ Xu dWu |p )
0 s≤u s≤t 0
2. Next, we compute
Z t Z t
∇Xt − δ Xt = b̃u (∇Xu − δ Xu )du + σ̃u (∇Xu − δ Xu )dWu + Rt
0 0
with
Z t Z t
Rt = ∇Xu (b0 (Xux ) − b̃u )du + ∇Xu (σ 0 (Xux ) − σ̃u )dWu .
0 0
And we compute
" # Z
T
2 0 0
E sup |Rt | ≤ CE |∇Xu (b (Xux ) − b̃u )|2 + |∇Xu (σ (Xux ) − σ̃u )|2 du .
t≤T 0
58
We observe that
Z 1
0
|b (Xux ) − b̃u | =| (b0 (Xux ) − b0 (Xux+ + λ∆ Xu ))dλ| ≤ C|∆ Xu |,
0
leading to
Z " #1
T 2
0
E |∇Xu (b (Xux ) − b̃u )|2 du ≤ CE sup |∆ Xu | 4
,
0 u≤T
Z " #1
T 2
Z t
∇Xtx =x+ σ 0 (Xsx )∇Xsx dWs ,
0
Z t
∇Xtx+ = x + + σ 0 (Xsx+ )∇Xsx+ dWs .
0
Z u Z u
0
Γu =+ σ (Xsx )Γs dWs + σ 0 (Xsx ) − σ 0 (Xsx+ ) ∇Xsx+ dWs .
0 0
Z u Z u
0
sup |Γu |2 2
≤ C( + sup | σ (Xsx )Γs dWs |2 + sup | σ 0 (Xsx ) − σ 0 (Xsx+ ) ∇Xsx+ dWs |2 )
u≤t u≤t 0 u 0
Z t Z t
2 2 2 x x+ 2 x+ 2
E sup |Γu | ≤ C( + E |Γs | ds + |Xs − Xs | |∇Xs | ds )
u≤t 0 0
59
Using CS inequality, we get
Z t " #1
2
Z t Z t
0
∇Xtx =1+ b (Xsx )∇Xsx ds + σ 0 (Xsx )∇Xsx dWs .
0 0
leading to
0 x 2
0 (X x )− σ (Xt ) )ds+ t
Rt
σ 0 (Xtx )dWs
R
∇Xtx = e 0 (b s 2 0 .
Example 3.2. Set d = 1. Give the expression of the tangent process in the BS
model.
60
3.2.2 Computing the delta (pathwise approach)
Proposition 3.2. Assume that g is Cb1 and b and σ are Cb2 , then u is Cb1 and
Proof. For the proof, we assume that g is Cb2 . We are going to compute the limit
of ∆F () for −→ 0.
1. We have that
Z 1
1 x+ x
0 x ∆ XT
∆F () = E g(XT ) − g(XT ) = E g (XT + λ∆ XT ) dλ
0
We then compute
Z 1
0 ∆ XT 0 0
∆ XT
|E[∇g(XTx )∇XTx ] x
− ∆F ()| = |E g (XT ) ∇XT −x
− x x
g (XT + λ∆ XT ) − g (XT ) dλ |
0
x ∆ XT |∆ XT |2
≤ CE |∇XT − | + CE
∆ XT
≤ C(E |∇XTx − | + )
using (3.2). Using (3.1), we then obtain lim→0 ∆F () = E[∇g(XTx )∇XTx ], proving
the differentiability f u.
|∂x u(0, x + ) − ∂x u(0, x)| = |E g 0 (XTx+ )∇XTx+ − g 0 (XTx )∇XTx |
≤ |E g 0 (XTx+ )(∇XTx+ − ∇XT2 ) | + |E (g 0 (XTx+ ) − g 0 (XTx ))∇XT2 |
1
≤ CE |∇XTx+ − ∇XTx | + CE |XTx+ − XTx |2 2 ,
where for the last term we used CS inequality and Lemma 3.1. Now using (3.3)-(3.2),
we have
61
Letting goes to 0 concludes the proof. 2
62
3.2.3 Practical implementation
It requires the discretization of (X x , ∇X x ). One may use the usual Euler scheme,
setting:
- and on π:
d
X
Xtπi+1 = Xtπi + b(Xtπi )(ti+1 − ti ) + σ .j (Xtπi )(Wtji+1 − Wtji )
j=1
d
X
∇Xtπi+1 = ∇Xtπi + ∇b(Xtπi )∇Xtπi (ti+1 − ti ) + ∇σ .j (Xtπi )∇Xtπi (Wtji+1 − Wtji )
j=1
Remark: simplification if d = 1.
63
3.3 Greek weights
h i
u(θ) = E g(XTθ )
Example 3.3. (i) If θ = x, and x is the starting point of X then we are computing
the ’Delta’.
the ’vega’.
h i
∂θ u(θ0 ) = E g(X θ0 )Π
• Precisely, we consider
h i
2 θ0
W := {Π ∈ L | ∂θ u(θ0 ) = E g(X )Π , for all bounded measurable function g}
h i
∂θ u(θ0 ) = E g(X θ0 )Π + E()
error.
64
3.3.1 Likelihood Ratio Method
• Assuming that XTθ has a differentiable positive probability density φ(θ, x):
h i h i
∂θ E g(XTθ ) = E g(XTθ0 )s(θ0 , XTθ0 ) with s(θ, x) = ∂θ ln(φ(θ, x)) .
θ=θ0
We compute
h i Z
∂θ E g(XTθ ) = ∂θ g(x)φ(θ, x)dx
Z Z
∂θ φ(θ, x)
= g(x)∂θ φ(θ, x)dx = g(x) φ(θ, x)dx
φ(θ, x)
W := {Π ∈ L2 | E Π | XT0 = S 0 }
V ar[g(XT0 )Π] = E (g(XT0 )Π)2 − (∂θ u)2
= E E (g(XT0 )Π)2 |XT0 − (∂θ u)2
≥ E (E g(XT0 )Π|XT0 )2 − (∂θ u)2
= V ar[g(XT0 )S 0 ]
65
Example: Black Scholes delta In the Black-Scholes setting, we have
−rT x WT
∂x u(0, x) = e E g(Xt ) .
xσT
1 log(y/x) − (r − 21 σ 2 )T
h(y) = √ ϕ(ζ(y)), ζ(y) = √
yσ T σ T
And we compute
log(y/x) − (r − 12 σ 2 )T
∂x h(y)/h(y) = −ζ(y)∂x ζ(y) =
xσ 2 T
2
−rT x WT 1
∂σ u(0, x) = e E g(Xt ) − WT − .
σT σ
Higher order derivatives Assuming that XTθ has a twice differentiable positive
h i 2 θ
2 θ ∂θθ φ(θ, XT )
∂θθ E g(XTθ ) = E g(XT ) .
φ(θ, XTθ )
2
2 −rT 1 x WT 1
∂xx u(0, x) =e E g(Xt ) − WT − .
x2 σT σT σ
66
3.3.2 Integration by part
• We consider the delta in the Black Scholes setting (r = 0), the payoff function
is Cb1 .
1 0 x x
∂x u(0, x) = E g (XT )XT .
x
1
∂x u(0, x) = E[g(XTx )WT ] .
xσT
We have
Z √ √
2 T /2+σ 2 T /2+σ
∂x u(0, x) = g 0 (xe−σ Ty
)e−σ Ty
ϕ(y)dy
which rewrites
Z √
1 2 T /2+σ
∂x u(0, x) = √ ∂y [g(xe−σ Ty
)]ϕ(y)dy
xσ T
Z √
1 2 T /2+σ
= √ g(xe−σ Ty
)yϕ(y)dy
xσ T
67
3.3.3 Bismut’s formula
Theorem 3.2.
Z T
x 1 x −1 x
∂x u(0, x) = E g(XT ) σ(Xs ) ∇Xs dWs . (3.4)
T 0
1
RT
(Here we have Π = T 0 σ(Xsx )−1 ∇Xsx dWs ).
and then
Z T
1
∂x u(0, x) = E[∂x u(s, Xsx )∇Xsx ] ds
T 0
Z T Z T Z T
E ∂x u(s, Xsx )∇Xsx ds =E ∂x u(s, Xsx )σ(Xsx )dWs σ(Xsx )−1 ∇Xsx dWs
0 0 0
And via the martingale representation theorem, since u is solution to the pde,
Z T
∂x u(s, Xsx )σ(Xsx )dWs = g(XTx ) − u(0, x)
0
Wt
∂x E[φ(Xt )] = E φ(Xt ) . (3.5)
σt
68
4 U.S. options in complete market
(σ is invertible).
Upon exercise, the option pays to the owner g(Xt ) (g exercise payoff, Lipstchitz
continuous).
• Super-hedging price:
y,φ
pus
0 := inf{y ∈ R | ∃ φ; admissible financial strategy s.t. Vt ≥ g(Xt ) ∀ t ∈ [0, T ]}
,→ at time 0:
−rτ
pus
0 = sup E e g(Xτ )
τ ∈T[0,T ]
h ?
i
= E e−rτ g(Xτ ? )
,→ at time t ∈ [0, T ]:
h i
−r(τ −t)
pus
t = esssupτ ∈T[t,T ] E e g(Xτ )|F t
69
h i
• (non-linear) PDE representation: pus (t, x) = supτ ∈T[t,T ] E e−r(τ −t) g(Xτt,x ) is
solution3 to
min{−LX u − ru , u − g} = 0 on [0, T ) × R
u(T, .) = g(.)
• This comes from the Dynamic programming principle: For all stopping time
t≤θ≤T
h i
pus (t, x) = sup E e−r(τ −t) g(Xθt,x )1{θ>τ } + pus (θ, Xθt,x )1{θ≤τ }
τ ∈T[t,T ]
3
in some sense... e.g. viscosity sense.
70
4.2 Bermudan option
• An option that can be exercised at a discrete set of time (<) during its life.
< = {0 =: s0 , . . . , sj , . . . , sκ := T } .
• Super-hedging price:
,→ at time 0:
pb0 = sup E e−rτ g(Xτ )
<
τ ∈T[0,T ]
h ∗
i
= E e−rτ g(Xτ ∗ )
,→ at time t ∈ <:
h i
pbt = esssupτ ∈T < E e−r(τ −t) g(Xτ )|Ft
[t,T ]
h i
−r(sj+1 −sj )
Ysj = max Csj , Gsj where Csj := E e Ysj+1 |Fsj (Continuation value)
71
Proposition 4.1. (i) Y is the smallest super-martingale above G = g(X).
τ ? = inf{t ∈ < | Yt = Gt } ∧ T ,
i.e. Y0 = E[Gτ ? ].
Proof.
(i) by induction
< .
for all τ ∈ T[0,T ]
=0
(iii) For t ∈ <, we have that Yt = Y0 + Mt? − A?t where M ? is a martingale and A?
Ms?j+1 = Ms?j + Ysj+1 − E Ysj+1 |Fsj
A?sj+1 = A?sj + Ysj − E Ysj+1 |Fti
72
We notice that Ysj − E Ysj+1 |Fti = [Gsj − Csj ]+
(iv) Denotes
y = V0y,φ ≥ Y0
Y0 = E[ητ?? ] == E[ηT ] .
In our setting, this means that we can replicate ηT with an initial wealth of Y0 i.e.
Thus Y0 ∈ Γ. 2
73
Proposition 4.2. Set |<| = max0<j≤κ (sj − sj−1 ), then the following holds:
h i
sup E |pbt − pus
t |2
≤ C|<|α
t∈[0,T ]
0 ≤ pus b
t − pt
Let τ̄ , be the projection on the grid < of τ ∗ (optimal stopping time for the US
option), we have
h i
pus
t − pb
t = E g(Xτ ∗ ) Ft − p
b
t
h i
≤ E[g(Xτ ∗ )|Ft ] − E g(Xτ̄ ) Ft
h i
≤ CE |Xτ ∗ − Xτ̄ | Ft
p
≤C |<|
Z τ̄
pus
t − pbt ≤ CE | LX g(Xs )ds| Ft
τ∗
≤ C|<|
74
4.2.1 Discretisation of the forward process
pπ0 = sup E e−rτ g(Xτπ )
<
τ ∈T[0,T ]
h ∗
i
= E e−rτ g(Xτπ∗ )
1
|pb0 − pπ0 | ≤ C|π| 2 .
75
Extension to the continuous case
√
|pus π
0 − p0 | ≤ C π .
76
4.2.2 Longstaff-Schwarz algorithm
• Observe that in Definition (4.1), the optimal stopping time τ̂0 (from time 0)
can be estimated by
1. set τ̂κ = T
execution
1. set τ̃κ = T
2. then set τ̃j = sj 1Ej + τ̃j+1 1Ejc with Ej = {E Gτ̃j+1 |Fsj ≤ Gsj } .
• Note that the value process Y is computed only through its representation in
indeed, Ysj = E Gτ̃j+1 |Fsj
77
4.3 Dual approach
The inf is achieved for M ? the martingale part of the Doob-Meyer decomposition
and
Y0 = sup(Gt − Mt? ) .
t∈<
• The previous representation has been introduced in [71] (for the continuous
Y0 = sup E[Gτ − Mτ ]
τ
≤ E sup(Gτ − Mτ ) = E sup(Gt − Mt ) .
τ t∈<
Gt ≤ Yt = Y0 + Mt? − A?t
which leads to
78
• In [71], sub-optimal martingales are constructed in "ad-hoc" way, see Section
• Numerical methods based on the dual formula 4.1 will lead naturally to upper-
79
4.4 Implementation using regression techniques
< = {0 =: s0 , . . . , sj , . . . , sκ := T }
Yj = max E Yj+1 |Fsj , g(Xsj )
• Denote Cj := cj (Xsj ) := E Yj+1 |Xsj = E Yj+1 | Fsj ( continuation value),
we have
Yj = cj (Xsj ) ∨ g(Xsj )
• Observe that
Cj = argminZ∈L2 (Fsj ) E |Yj+1 − Z|2 = argminZ∈L2 (σ(Xsj )) E |Yj+1 − Z|2
80
• Denoting L2j the set of PXsj -square integrable function, we observe
cj = argminf ∈L2 E |Yj+1 − f (Xsj )|2
j
• L2j is too big! so one considers some basis function (ψ` )`≥1 and then pick a
K
X
f' β ` ψ`
`=1
K
X
c̄j := β̄j` ψ` , and cj = c̄j + error
`=1
PK `
(Cj = C̄j + error where C̄j = `=1 c̄j ψ` (Xsj ))
• The coefficient β̄ is easily calculated, once observed that C̄j is the orthogonal
h i h i
Bψj = E ψ(Xj )> ψ(Xj ) , j
Bψu = E ψ(Xj )> uj+1 (Xj+1 ) ,
we have that
81
By property of the orthogonal projection on the vector space V, we have
E (Yj+1 − C̄j )Z = 0, for all Z ∈ V leading to
" K
#
X
E Yj+1 − β̄` ψ` (Xj ) ψr (Xj ) = 0 , 1 ≤ r ≤ K .
`=1
j
• In practice, one has to consider estimated counterparts of Bψj , Bψu ...
N
1 X
Bψj = ψ(Xji )> ψ(Xji )
N
i=1
82
• Full approximation (Tsitsiklis-Van Roy): observe that in practice one has to
PK ˆ
set c̄ˆj := `=1 β̄` ψ` and ˆj = c̄ˆj ∨ g.
ū
• One can compute an approximated optimal policy τ̂ ∗ on each path and recom-
Remark 4.1. 1. The choice of the basis functions is key, specially in high di-
mension.
• Based on Theorem 4.1, one should find a "good" martingale. Various ap-
proaches beyond "guessing" [71] have been considered see e.g. [8] and the
references therein.
∆Mj? = Ysj+1 − E Ysj+1 |Fsj . (4.3)
83
• To ensure the martingale property, Haugh & Kogan suggest to recompute the
– Simulate M path of X: for each Xsmj , compute E ū
ˆ(Xsj+1 )|Xsj by res-
ˆ j with ∆
M̂j+1 = M̂j + ∆ ˆ j = ū
ˆ(Xsj+1 ) − E ū
ˆ(Xsj+1 )|Xsj (4.4)
1 PM m
– Compute M m=1 maxt∈< (g(Xsj ) − M̂jm )
84
4.5 Quantization based methods
See e.g. the review article [62] or Chapter 5 in the book [63]
{x1 , . . . , xM }.
s.t.
Ci (x) ⊂ {ξ ∈ Rd | |ξ − xi | ≤ min |ξ − xj |}
i6=j
85
• x-quantization of X: X̂ = Px (X) (remark: if X is absolutely continuous any
h i1
2
• Quadratic mean quantization error : E |X − X̂ x |2 =: kX − X̂ x k2
h i
1. The quadratic distortion function: x 7→ E |X − X̂ x |2 reaches a minimum
at some quantizer x∗
h ∗
i
2. The function M 7→ E |X − X̂ x |2 is decreasing to 0 as M → +∞
• Upper bound on the convergence rate for X ∈ L2+ for some > 0: There
1
∀M ≥ 1, min kX − X̂ x k2 ≤ Cd, kXk2+ M − d (4.5)
x∈(Rd )M
∗
kX − X̂ x k2 := min{kX − Y k2 | #Im(Y ) ≤ M } (4.6)
min |X − yi | ≤ |X − Y | P − a.s.
i
∗
so that kX − Y k2 ≥ kX − X̂ y k2 ≥ kX − X̂ x k2 . 2
86
• Any L2 -optimal M -quantizer x∗ ∈ Rd is stationary in the following sense
h ∗
i ∗
E X X̂ x = X̂ x . (4.7)
h i
(in particular E[X] = E X̂ x as soon as the quantizer is stationary which
Proof. We compute
h ∗
i
E |X − X̂ x |2
h i2 h i h i h i
2
x∗ x∗ x∗ x∗ x∗ x∗
= E X − E X X̂ + 2(X − E X X̂ )(E X X̂ − X̂ ) + E X X̂ − X̂
h i2 h i
∗ ∗ ∗ 2
= E X − E X X̂ x + E X X̂ x − X̂ x
∗
where the last inequality is obtained by conditioning w.r.t X̂ x .
h ∗
i
From the previous point, we also have, as E X X̂ x takes almost M different
∗
values (this is a measurable function of X x ...)
h i h i 2
x∗ 2 x∗
E |X − X̂ | ≤ E X − E X X̂
This leads to
h i
2
x∗ x∗
E E X X̂ − X̂ =0.
• Two M -quantizer (M = 500) of N (0, I2 ) one of them being (close to be) op-
timal... [62]
87
For a discussion on how to obtain optimal quantization grid, we refer to Section
5.3 in [63].
88
Cubature formulas
h i XM
E[g(X)] ' E g(X̂ x ) = f (xi )P(X ∈ Ci (x))
i=1
• If g is Lipschitz,
h i h i
|E[g(X)] − E g(X̂ x ) | ≤ [g]Lip E |X − X̂ x | ≤ [g]Lip kX − X̂ x k2
h i h i
|E[g(X)] − E g(X̂ x ) | ≤ [Dg]Lip E |X − X̂ x |2 (4.8)
as X̂ x is stationnary.
Conditional expectation
h i
• Approximation of Θ = E[F (X)|Y ] by Θ̂ = E F (X̂)|Ŷ ...
kΘ − Θ̂k2 ≤ C kX − X̂k2 + kY − Ŷ k2 (4.9)
89
4.5.2 Quantization tree for optimal stopping problem
This technique has been introduced in [6, 5], where a complete error analysis is done,
Ck := (Cki )1≤i≤Mk .
h i
• For any ϕ, we replace E[ϕ(Xk+1 )|Xk ] by E ϕ(X̂k+1 ) | X̂k , denote then
k
πij = P(X̂k+1 = xjk+1 |X̂k = xik )
Pn−1
• ’online’ computational cost: ∼ k=0 Mk Mk+1 .
90
• note that the grids and (π k ) are computed offline, e.g. by MC method: Note
k
]{n|X̂kn = xik & X̂k+1
n = xjk+1 }
πij = limN →∞
]{n|X̂kn = xik }
where (Xkn )1≤n≤N,1≤k≤κ are MC simulation of the Markov Chain (Xk )1≤k≤κ .
91
4.5.3 Markovian quantization (grid method)
for x ∈ Rd ,
δbδ −1 (xj − X0j ) + 21 c + X0j , if |xj − X0j | 6 κδ,
(Π[x])j = κδ, if xj − X0j > κδ,
−κδ, if xj − X0j < κδ.
p ∆Wi
ci :=
∆W hi GM ( √ )
hi
GM denotes the projection operator on the optimal quantization grid for the
h i1 √ 1
c
E |∆Wi − ∆Wi |p p
6 Cp,d hM − d . (4.12)
92
Definition 4.3. We denote (Ybiπ )06i6n the solution of the backward scheme satisfying
h i
Ybiπ = max(Eti Ybi+1
π biπ ))
, g(X (4.14)
Proposition 4.4. For all i ∈ {0, ..., n}, there exists a function uπ (ti , .) : Γ → R
such that
Yb π = uπ (ti , X
biπ )
This function is computed on the grid by the following backward induction: for all
h p i
uπ (ti , x) = max{E uπ ti+1 , Π x + hi b(x) + hi σ(x)GM (U ) , g(x)}
93
5 Non-linear pricing methods
• We work in a Lipschitz setting. We give the main definition and properties for
solution to BSDEs.
• first introduced by Bismut [10, 11] and then studied in general way first by
5.1.1 Definition
for some progressively measurable random function f , called the driver, and a
• We denote S 2 (Rk ) the vector space of RCLL5 adapted processes Y , with values
94
where for z ∈ Rk×d , |z|2 = Tr(zz † ).
valued in Rk is such that for all (y, z) ∈ Rk × Rk×d , the process {f (t, y, z)}0≤t≤T is
f (t, y, z) − f (t, y 0 , z 0 ) ≤ L |y − y 0 | + kz − z 0 k ;
2. Integrability condition:
Z T
2 2
E |ξ| + |f (r, 0, 0)| dr < ∞.
0
Theorem 5.1. Under (H1), there exists a unique solution (Y, Z) ∈ Sc2 × H2 to
Z T Z T
Yt = ξ + f (s, Ys , Zs )ds − Zs dWs , 0 ≤ t ≤ T. (5.3)
t t
Linear BSDEs We first study linear BSDE for which we can give an almost
explicit solution. For this section, we set k = 1: Y is then real valued and Z a
Proposition 5.1. Let {(at , bt )}t∈[0,T ] be progressively measurable and bounded pro-
cesses with value in R × Rd . Let {ct }t∈[0,T ] be an element of H2 (R) and ξ a random
95
The linear BSDE
Z T Z T
Yt = ξ + {ar Yr + Zr br + cr } dr − Zr dWr , (5.4)
t t
Z T
∀t ∈ [0, T ], Yt = Γ−1
t E ξΓT + cr Γr dr Ft , (5.5)
t
nZ t
1
Z t Z t o
2
Γt = exp br dWr − |br | dr + ar dr .
0 2 0 0
Comparison Theorem
Theorem 5.2. Let k = 1 and assume that (ξ, f ) satisfies (H1), the solution to
the associated BSDE is denoted (Y, Z). Let (Y 0 , Z 0 ) be a solution of a BSDE with
RT
parameters (ξ 0 , f 0 ) and satisfying 0 f 0 (t, Yt0 , Zt0 )dt ∈ L2 (FT ). We also assume that
measure). Then,
96
5.1.3 Application to non-linear pricing
and assume a different rate for lending (r) and borrowing (R, R > r).
under Q ∼ P s.t.
Z t Z t
Xt = x + rXs ds + σ(s, Xs )dWs (5.7)
0 0
• A portfolio is constituted of cash α and risky asset (quantity φ), its value at
time t is
Vtυ,φ = αt + φt Xt . (5.8)
- the cash account is: [αt ]+ rdt − [αt ]− Rdt = (rαt + [αt ]− (r − R))dt
g(XT ) is defined as
97
equivalently one has to solve for (Y, ∆) the following
Z T Z T
−
Yt = g(XT ) + (rYs + [Ys − ∆s Xs ] (r − R))ds − ∆s σ(s, Xs )dWs
t t
(5.11)
Z T Z T
Zs
Yt = g(XT ) + (rYs + [Ys − Xs ]− (r − R))ds − Zs dWs (5.12)
t σ(Xs ) t
Setting
Zs
φs = (5.13)
σ(s, Xs )
Now let v s.t. ∃ψ, VTv,ψ ≥ g(XT ) = YT . We observe that V v,ψ has the same
dynamics as a BSDE and thus we can use the comparison theorem to conclude that
98
5.2 Main properties in the Markov setting
See [66]
− u(0) (t, x) − f x, u(t, x), ∂x u(t, x)σ(t, x) = 0 (t, x) ∈ [0, T ) × Rd (5.14)
• Assuming that u is a smooth solution to the above non-linear pde, one can
prove conversely
define
For a multi-index, β ∪n≥1 {0, 1}n with β = (j1 , . . . , jk ) for some k ≥ 1, we denote
by −β := (j2 , . . . , jk ) and
ϕβ = Lβ ϕ = L(j1 ) [L−β ϕ] .
99
5.3 Numerical analysis of backward Methods
- at tn = T :
- i < n, compute
h i
π ( Wti+1 −Wti )
Ziπ = Eti Yi+1 ti+1 −ti
(5.20)
π
Yiπ = Eti Yi+1 + (ti+1 − ti )f (Yiπ , Ziπ )
Remark 5.2. .
1
(i) This is an implicit Euler scheme and the error is O(|π| 2 ) in the Lipschitz case
and assuming smoothness of the coefficient O(|π|) at most, generally (see below).
(ii) Explicit version of the scheme: in the smooth case Znπ := σ(XT )g 0 (XT ).
heuristics.
100
5.3.1 L2 -stability
h i
Ỹi = Eti Ỹi+1 + (ti+1 − ti )f (Ỹi , Z̃i ) + ζi (5.21)
Wti+1 − Wti
Z̃i = Eti Ỹi+1 ( ) (5.22)
ti+1 − ti
where ζi ∈ L2 (Fti ).
Definition 5.1 (L2 -stability). The scheme given in (5.20) is L2 -stable if there exists
Yiπ = Yi+1
π
+ hi f (Yiπ , Ziπ ) − hi Ziπ Hi − ∆Mi (5.24)
∆Wi
where Hi = h . Note that (5.26) defines ∆Mi , moreover it satisfies
Eti[∆Mi ] = Eti[∆Mi Hi ] = 0 and E |∆Mi |2 < ∞ (5.25)
These properties are obtained by using the definition of the scheme given in (5.20).
101
Denoting δfi = f (Yiπ , Ziπ ) − f (Ỹi , Z̃i ) and δ∆Mi = ∆Mi − ∆M̃i , we observe that
Squaring both sides and taking conditional expectation, we compute, using Young’s
inequality,
h i C
|δYi |2 + hi |δZi |2 ≤ (1 + Ch)Eti (δYi+1 + hi δfi )2 + |ζi |2 . (5.28)
h
Note that
h
≤ (1 + )(|δYi+1 | + Chi |δYi |)2 + C(1 + )h2i |δZi |2 (5.30)
h
1
(δYi+1 + hi δfi )2 ≤ (1 + Ch)|δYi+1 |2 + Ch|δYi |2 + h2i |δZi |2 (5.31)
2
C
|δYi |2 ≤ (1 + Ch)Eti |δYi+1 |2 + |ζi |2 (5.32)
h
C
≤ eCh Eti |δYi+1 |2 + |ζi |2 (5.33)
h
102
5.4 Convergence analysis assuming no error on X
• Let us introduce
∆Wi
Ẑti = Eti Yti+1 . (5.34)
hi
Z ti+1
ζ̂i := Eti [f (Yt , Zt ) − f (Yti , Ẑti )]dt . (5.35)
ti
X h i
T (π) := E n|ζ̂i |2 . (5.36)
i
• Assume at this point that there is no error made on the forward process, thus
max E |Yti − Yiπ |2 ≤ CT (π) . (5.37)
i
Proof. This comes directly from the L2 -stability of the scheme with the
• we now study the order of the truncation error under (HL) or (Hr).
103
5.4.2 Order of convergence in the smooth case
h i
|Eti[f (Yt , Zt ) − f (Yti , Zti )] | = |Eti u(0) (t, Xt ) − u(0) (ti , Xti ) | (5.40)
Z ti+1
(0,0)
= |Eti u (t, Xt )dt | (5.41)
ti
≤ C|π| (5.42)
where we used Ito’s formula for the last equality. Now we compute, setting Hi :=
∆Wi
hi
Z ti+1 Z ti+1
(0) 1 (1)
Ẑti = Eti u(ti+1 , Xti+1 )Hi = Eti Hi u (t, Xt )dt + u (t, Xt )dt
ti hi ti
(5.43)
Observe that
Z ti+1 Z ti+1
|Eti Hi u(0) (t, Xt )dt | = |Eti Hi {u(0) (t, Xt ) − u(0) (ti , Xti )}dt | (5.44)
ti ti
≤ C|π| (5.45)
and that
h i Z t
(1) (1) (0,1)
Eti u (t, Xt ) = Eti u (ti , Xti ) + u (s, Xs )ds (5.46)
ti
104
We thus get
Z ti+1
1 (1)
|Eti u (t, Xt )dt − u(1) (ti , Xti )| ≤ C|π| (5.47)
hi ti
And, by summing this local error estimate, we conclude that T (π) ≤ C|π|2 . 2
105
5.4.3 Order of convergence in the Lipschitz case
Z !
2
n−1
X ti+1
T (|π|) ≤ C |π| + max E |Yt − Yti | + E |Zt − Z̄ti |2 dt (5.49)
i ti
i=0
where
Z ti+1
1
Z̄ti = Eti Zt dt .
hi ti
leading to
Z ti+1
2 2
|Z̄ti − Ẑti | ≤ CEti |f (Yt , Zt )| dt (5.53)
ti
Then
Z ti+1 Z ti+1
2 2 2 2
|Eti [f (Yt , Zt ) − f (Yti , Ẑti )]dt | ≤ hi Eti C [|Yt − Yti | + |Zt − Z̄ti | + |Z̄ti − Ẑti | ]dt
ti ti
(5.54)
Z ti+1
2 2
≤ hi CEti [|Yt − Yti | + |Zt − Z̄ti | ]dt
ti
(5.55)
Z ti+1
+ Ch2i Eti 2
|f (Yt , Zt )| dt (5.56)
ti
106
The proof is concluding by summing on i and recalling that
Z T
2
E |f (Yt , Zt )| dt ≤ C. (5.57)
0
Z ti+1
X Z ti+1
2 2
max E |Zt | dt + E |Zt − Z̄ti | dt ≤ C|π| (5.58)
i ti ti
i
≤ C|π| (5.61)
where we used (5.58). Then, the proof is concluded by combining the previous
107
5.5 Full discrete-time error analysis
• The scheme for (Y, Z) is the same but the terminal condition (5.19) is now
∆Wi
Z̃i = Eti Ỹi .
ti+1 − ti
h i
Ỹi = Eti Ỹi+1 + (ti+1 − ti )f (Ỹi , Z̃i ) + ζi (5.64)
where
h i
ζi = Eti ζie + ζif + ζiz (5.65)
and ζie , ζif and ζiz are defined in (5.67), (5.69) and (5.70) respectively.
Proof. (b = 0) 1. Applying Ito’s formula, we have, setting Vsi = σ(Xtπi )∂x u(s, Xsπ )
Z ti+1 Z ti+1
1
u(ti+1 , Xtπi+1 ) = u(ti , Xtπi ) + {∂t u(s, Xsπ ) + σ 2 (Xtπi )∂xx
2
u(s, Xsπ )}ds + Vsi dWs
ti 2 ti
(5.66)
108
Then, introducing
Z ti+1
1 2
ζie := − σ 2 (Xtπi ) − σ 2 (Xsπ ) ∂xx u(s, Xsπ )ds (5.67)
2 ti
Z ti+1 Z ti+1
Ỹi+1 = Ỹi − hf (Ỹi , Vtii ) + u(0) (s, Xsπ ) − u(0) (ti , Xtπi ) ds + ζie + Vsi dWs
ti ti
(5.68)
Now we define
Z ti+1
ζif =− u(0) (s, Xsπ ) − u(0) (ti , Xtπi ) ds (5.69)
ti
h i
Ỹi = Eti Ỹi+1 + hf (Ỹi , Z̃tii ) + ζie + ζif + ζiz . (5.71)
109
Proof. See Exercise II.22 2
We simply observe that (Ỹi , Z̃i )i defines a pertubed scheme as in Section 5.3.1. We
can then combine Proposition 5.2 with Lemma 5.1 to conclude the proof, recall also
(5.37).
110
5.6 Numerical illustration and further consideration
• Regression methods, as the one used for US options, can be easily adapted to
the BSDEs setting, introduced in [37, 56] and extensively studied in [39, 40,
38, 34]
ticular to work with degenerate diffusion), see e.g. [27, 28, 26].
Z 1 Z 1
ω(1, W1 ) 5
Yt = + (Zs1 + Zs2 + Zs3 )( − Ys )ds − Zs dWs . (5.75)
1 + ω(1, W1 ) t 6 t
where ω(t, x) = exp (x1 + x2 + x3 + t). Applying Itô formula, we verify that the
solution is given by
ω(t, Wt ) ω(t, Wt )
Yt = and Ztl = , l ∈ {1, 2, 3} , t ∈ [0, 1] . (5.76)
1 + ω(t, Wt ) (1 + ω(t, Wt ))2
• In this very smooth setting, we can introduce higher order scheme too. We
111
• The number of points used to approximate ∆W depends on the theoretical rate
112
5.6.3 (Markovian) quantisation
• We illustrate here the Markovian quantisation with Example [23] in the case
of quadratic BSDEs.
• "Classical" quantization can be used for BSDEs as well, see e.g. [64].
where a, ν, and (X0` )`∈{1,2,3} are given real positive parameters and g : Rd → R is a
Applying Ito’s formula, one can show that the solution is given by
1
Yt = log Et exp ag(X1 ) , t61. (5.78)
a
For any given g, ν and a, it is possible to estimate the solution Y0 at time 0 using an
ν2
+νW1`
approximation of the Gaussian distribution at time T = 1, since X1` = X0` e− 2 .
2
X
g : x 7→ 3 sin2 (x` ) , (5.79)
`=1
• The non-Lipschitz setting may cause instability and we use in the graph below
113
Figure 8: Comparison of schemes’ convergence
Z t Z t
Yty,Z =y− f (Ysy,Z , Zs )ds + Zs dWs (5.80)
0 0
h i
V := min E |g(XT ) − YTy,Z |2 (5.81)
(y,Z)∈R×H2
• In the Lipschitz setting, the solution of the above optimisation is the BSDEs
(5.1) with V = 0!
• Main idea: solve numerically the optimisation problem (5.81) to get an ap-
114
5.7.1 Discretization of the optimisation problem (5.81)
π:
Ytπn+1 = Ytπn + hf (Ytπn , Ztn ) + Ztn (Wtn+1 − Wtn ) and Y0π = y . (5.82)
specification!
the coefficient of a Neural Network ϕNN and ZtΘn = ϕNN (tn , Xtn )
K
X
ϕL (tn , ·) = θk φk (tn , ·) , θ ∈ RdK
k=1
V π := min E |g(XTπ ) − YTπ,υ |2 (5.83)
υ∈R1+K̄
specification.
115
6 An introduction to McKean-Vlasov SDEs in finance
See the seminal paper [72] and among others [42, 52], [20] for the large population
6.1 Definition
6.1.1 Introduction
N
X
1
dXti,N = Xtj,N − Xti,N dt + dBti (6.1)
N
j=1
LLN should lead to the following equation when the number of banks goes to infinity:
dX̄t = (E X̄t − X̄t )dt + dBt (6.2)
,→ Solution is given by an Ornstein-Uhlenbeck process (note E X̄t = E[X0 ]).
• Typical questions:
1. Does the solution of (6.2) describes well the system (6.1) when N → +∞?
116
6.1.2 Notations
R
• For p ≥ 1, Pp (Rd ) is the space of probability measure µ satisfying |x|p µ(dx) <
∞.
• Wasserstein distance on Pp :
1
Wp (µ, ν) := inf E[|X − Y |p ] p . (6.3)
X∼µ,Y ∼ν
Note that (Pp , Wp ) is a polish space6 see e.g [75] or [12] among others.
1 PN 1 PN
• For later use: denote µN = N i=1 δxi and ν N = N i=1 δyi , then
N
1 X
W22 (µN , ν N ) ≤ |xi − yi |2 (6.4)
N
i=1
6
A complete metric space.
117
6.1.3 Existence and uniqueness
• We consider, for ξ ∈ L2 ,
Z t Z t
Xt = ξ + b(Xs , L(Xs ))ds + σ(Xs , L(Xs ))dWs , t ≤ T. (6.5)
0 0
|b(x, µ) − b(x0 , µ0 )| + |σ(x, µ) − σ(x0 , µ0 )| ≤ L |x − x0 | + W2 (µ, µ0 ) (6.6)
1 2
∂t µt = −∂x {b(x, µt )µt } + ∂xx {σ 2 (x, µt )µt } (6.7)
2
118
Proof. of Theorem 6.1 [b = 0 for the proof] Let Φ : S 2 → S 2 given by
Z t
Xt = Φ(x)t := ξ + σ(xs , L(xs ))dWs (6.8)
0
Z r
2
sup |∆Xr | ≤ sup | δσs dWs |2 (6.9)
r≤t r≤t 0
E |δσs |2 ≤ 2L2 E |xs − x0s |2 + W22 (L(xs ), L(x0s ))
≤ 4L2 E |xs − x0s |2 (6.10)
Z t
2 2
E sup |∆Xr | ≤ CE |δσs | ds (6.11)
r≤t 0
Z t
2
E sup |∆Xr | ≤ C E |xs − x0s |2 ds (6.12)
r≤t 0
Z t
2 2
E sup |∆Xr | ≤ C E sup |∆xr | ds (6.13)
r≤t 0 r≤s
" #
(CT )k
E sup |Φ(k) (x)r − Φ(k) (x0 )r |2 ≤ E sup |∆xr |2 ds (6.14)
r≤t k! r≤T
119
...
120
Proof. of Proposition 6.1. Let φ ∈ Cc ([0, T ) × R) and apply Ito’s formula to get
Z T
1
E[φ(t, Xt )] = E[φ(0, X0 )] + E (∂t φ + b(Xs , µs )∂x φ + σ(Xs , µs )2 ∂xx
2
φ)(s, Xs )ds
0 2
(6.15)
Z T Z
1
0= {(∂t φ + b(x, µs )∂x φ + σ(x, µs )2 ∂xx
2
φ)(s, x)}µs (dx)ds (6.16)
0 R 2
Recall that if µt has a smooth density that me denote mt : We can use Integration
Z
E[b(Xs , µs )∂x φ(s, Xs )] = b(x, µs )∂x φ(s, x)ms (x)dx (6.17)
R
Z
=− φ(s, x)∂x {b(x, µs )ms (x)}dx . (6.18)
R
Z
dXt = β(Xt , y)µt (dy)dt + dWt , (6.19)
X0 = ξ ∼ µ ∈ P2 (6.20)
where β is a Lipschitz function. (Note that b(x, ν) = E[β(x, χ)] where χ ∼ ν.)
121
Proof. We observe that
b(x, ν) − b(x, ν 0 ) = E β(x, χ) − β(x, χ0 ) for any χ ∼ ν, χ0 ∼ ν 0 (6.21)
|b(x, ν) − b(x, ν 0 )| ≤ LE |χ − χ0 | (6.22)
1
≤ LE |χ − χ0 |2 2 (6.23)
where we used Cauchy-Schwartz for the last inequality. Note that χ and χ0 are
arbitrary so that
1
|b(x, ν) − b(x, ν 0 )| ≤ L inf0 0
E |χ − χ 0 2 2
| = LW2 (ν, ν 0 ) (6.24)
χ∼ν,χ ∼ν
122
• Associated particle system:
N
1 X
dXti = β(Xti , Xtj )dt + dWti , (6.25)
N
j=1
X0i = ξ i
where (ξ 1 , . . . , ξ N ) are iid with law µ0 and (W i ) are i.i.d. Brownian motion.
• Observe that
N N
1 X 1 X
β(x, Xtj ) = b(x, µN
t ) where µ N
t := δX j (empirical measure part. syst.).
N N t
j=1 j=1
(6.26)
N
1 X
µ̄N
t := δX̄ j (6.28)
N t
j=1
• From this, one deduces that the law of the first k particules (X i ) converges to the
123
Proof. Let δX i := X i − X̄ i , r ≤ T ,
Z r Z r
|δXri |2 =| {b(Xsi , µN
s ) − b(X̄si , µs )}ds|2 ≤T |b(Xsi , µN i 2
s ) − b(X̄s , µs )| ds
0 0
(6.30)
leading to
Z t
i 2
E sup |δXr | ≤ CT E |b(Xsi , µN i N 2 i N i 2
s ) − b(X̄s , µ̄s )| + |b(X̄s , µ̄s ) − b(X̄s , µs )| ds
r≤t 0
(6.31)
Now,
N
X
1
E |b(Xsi , µN i N 2
s ) − b(X̄s , µ̄s )| ≤ CE|Xsi − X̄si |2 + |Xsj − X̄sj |2
N
j=1
≤ Cϕ(s) . (6.32)
124
Moreover,
N
1 X
|b(X̄si , µ̄N i 2
s ) − b(X̄s , µs )| = | {β(X̄si , X̄sj ) − b(X̄si , µs )}|2 (6.33)
N
j=1
Denote Y j = β(X̄si , X̄sj ) − b(X̄si , µs ) and observe that E Y j |X̄ i = 0, for j 6= i and
We then compute
2
XN
1
E |b(X̄si , µ̄N i
s ) − b(X̄s , µs )|
2
= 2 E Y j (6.34)
N
j=1
1 X h i X h i
= 2 E |Y i |2 + 2 E Y iY k + E Y jY k
N
k6=i j6=i,k6=i
N
X
1 C
= E |Y j |2 ≤ (6.35)
N2 N
j=1
Z t
C
ϕ(t) ≤ ϕ(s)ds + (6.36)
0 N
125
• It is possible also to obtain a weak error as follows
1
|E Ψ(µN
T ) − Ψ(µT )| = O( ) (6.37)
N
• Full discretization of the particles system requires a time discretization see e.g.
[14].
Lipschitz setting.
• Other type of approximation schemes are possible: via “projection” [9], via
126
6.3 Singular interaction
ν2 2
∂t u + u∂x u = ∂ u and u(0, ·) = g(·) (6.38)
2 xx
conservation law)
A simple model from fluid dynamics but represent also Carbon Allowance Price in
or alternatively
h i
E Ĝ0ν (x + νWt )
u(t, x) = −ν 2 h i (6.40)
E Ĝν (x + νWt )
G(y)
with Ĝν (y) = e− ν2 .
−w
Define θ := e η and compute
θ θ 2 2 (∂x2 w)2 θ
∂t θ = −∂t w , ∂x θ = −∂x w , ∂xx θ = − ∂xx w − (6.42)
η η η η
ν2 2 G(·)
∂t θ = ∂xx θ and θ(0, ·) = e− ν 2 (6.43)
2
127
This implies that
Z
1 G(y) |x−y|2
θ(t, x) = √ e− ν 2 − 2ν 2 t dy (6.44)
σ 2πt
Z
= Ĝν (x − y)φν (t, y)dy (6.45)
∂x θ
u = ∂x w = −ν 2 . (6.46)
θ
σ2 2
∂t v + ∂x (uv) = ∂ v and v0 = ∂x g (6.47)
2 xx
X0 ∼ v0 and
Z t
Xt = X0 + u(s, Xs )ds + σWt and Xt ∼ vt (6.48)
0
Rx R∞
We have u(t, x) = −∞ v(t, y)dy = −∞ H(x, y)v(t, y)dy with H(x, y) = 1y≤x .
• When ν = 0, there exists a unique "good physical solution" (a.k.a. entropy solution
x 0∧x∨t
u(t, x) = 1{0≤x≤t} + 1{x>t} = (6.49)
t t
M
X 1
dXti = H(Xti , Xtj )dt + σdWti . (6.50)
M
j=1
128
Legend. blue : σ = 1, orange: σ = 0.5, green: σ = 0.1, red: σ = 0
• The algorithm converges to the true value when σ → 0. Note that the sum is
obtained by sorting the system, so that denoting rk(Xti ) the rank of the particle Xti ,
rk(Xti )
Xtik+1 = Xtik + + σ∆Wki . (6.51)
M
M
X 1
dXti = (1 − H)(Xti , Xtj )dt + σdWti and X0 ∼ g 0 (6.52)
M
j=1
For g(x) = 1 ∧ (x − 1) ∨ 0, the algorithm behaves very well as it captures the shock
129
Legend. blue : T = 0, orange: T = 0.5, green: T = 1, red: T = 2
130
6.3.2 Calibration of LSV model
[45, 43]
gives
where C(T, K) is the price observed today (S0 = x) of Call option with strike K
and maturity T .
131
Stochastic volatility model .
132
Local Stochastic volatility model [57, 69].
• should incorporate volatility risk but also calibrate to European call option.
σD (t, x)
σ(t, x) = q (6.58)
E a2t |St = x
σD (t, St )at
dSt = rSt dt + q dWt (6.59)
E a2t |St
133
• This comes from Gyöngy Theorem [46], see also [18]. Namely, for an Ito process
b(t, x) = E[βt |Zt = x] and Σ(t, x)2 = E αt2 |Zt = x . (6.62)
134
Heuristics: 1. For Gyöngy Theorem (β = 0). Let φ ∈ Cc ([0, T ) × R) and apply Ito’s
formula to get
Z T
1 2 2
E[φ(t, Zt )] = E[φ(0, Z0 )] + E (∂t φ + αt ∂xx φ)(s, Zs )ds (6.63)
0 2
Z T
1 2 2
0=E (∂t φ + E αs |Zs ∂xx φ)(s, Zs )ds (6.64)
0 2
Z T
1 2
=E (∂t φ + Σ(s, Zs )∂xx φ)(s, Zs )ds (6.65)
0 2
Z T Z
1
0= (∂t φ + Σ(s, x)2 ∂xx
2
φ)(s, x)µs (dx)ds (6.66)
0 2
1 2
∂t m = ∂xx (Σ(t, x)2 m) and m0 = L(Z0 ) , (6.67)
2
2. Combining Gyöngy Theorem and Dupire result, we must have, recalling (6.57),
E (at σ(t, St ))2 |St = x = σD (t, x) (6.68)
• Existence and uniqueness for (6.59): mostly open, see nevertheless [1, 49] and [54].
135
Numerical simulation of (6.59) [77, 44] .
R
• Consider φ : R → R a smooth function s.t. φ(x)dx = 1 and set
1 x
φ (x) := φ( ) , ∀x ∈ R. (6.69)
N Z
1 X
µt (x) ' φ (x − St ) = φ (x − y)dµN
i
t (dy) (6.70)
N
i=1
- Nadaraya-Watson estimator at x
PN 2 i i
i=1 f (Yt )φ (x − St )
Θ(t, x) := E f 2 (Yt )|St = x ' PN =: ΘN (t, x) (6.71)
i
i=1 φ (x − St )
σD (t, x)
σ(t, x) ' σN (t, x) := p (6.72)
ΘN (t, x)
136
Heuristics.
1. Consider the kernel estimator of the density ρt (x, y) of L(Yt , St ) for two parame-
ters , 0 namely
N
1 X
φ (x − Sti )φ0 (y − Yti ) (6.73)
N
i=1
1 PN i
N i=1 φ (x − St )φ0 (y − Yti )
ρ(y|x) ' 1 P N
(6.74)
i
N i=1 φ (x − St )
and
1 PN i
R 2
i=1 φ (x − St ) f (y)φ0 (y − Yti )dy
E f 2 (Yt )|St = x ' N
1 P N
(6.75)
i
N i=1 φ (x − St )
Remark: if f 2 (y) = y and the function φ is symetric then one has directly
Z
yφ0 (y − Yti )dy = Yti . (6.76)
137
• Assume we are given an approximation Ȳ of Y on the grid {tk , k = 0, . . . , κ}.
• Acceleration in practice:
1. In σN (t,x), ΘN (t, x) is not computed for all x = Stjk but on a grid of R. The
2. The sums (6.71) has to be optimised for the method to be not too expensive
computation wise.
Remark 6.1. Note that the quantity E f 2 (Yt )|St can also be obtained by a non-
type:
f (Yt )
dXt = p dWt (6.80)
E[f 2 (Yt )|Xt ]
,→ the Markovian projection is indeed a Brownian motion and thus L(Xt ) = L(Wt ),
for all t.
and
20: (Legend: orange: true gaussian density, blue: estimation of XT density function)
138
• "Call price" in this model:
1. E[[X]+ ]:
True: 0.399
2. E[[X − 1]+ ]
3. E[[X − 0.5]+ ]
139
4. supremum:
140
Contents
I Handouts 15
1 Introduction 16
2.1.1 SDEs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
141
2.7 Acceleration methods . . . . . . . . . . . . . . . . . . . . . . . . . . . 53
3.1.1 Bias . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 61
3.1.2 Variance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 62
142
4.4 Implementation using regression techniques . . . . . . . . . . . . . . 87
143
5.6.1 Monte Carlo based methods . . . . . . . . . . . . . . . . . . . 121
II Exercises 158
144
Part II
Exercises
Exercise II.1. Prove inequality 2.4.
Exercise II.6. Let X π denote the Milstein scheme for X (one dimensional SDE)
given by
Z t Z t
Xt = X0 + b(Xs )ds + σ(Xs )dWs , t≤T,
0 0
for p ≥ 2 .
Z t Z t 21
2
φ(t) ≤ A φ(s)ds + B φ (s)ds +C , ∀t ≤ T ,
0 0
145
where A, B, C are positive constants.
2 )t
φ(t) ≤ 2Ce(2A+B .
1
π
Xi+1 = Xiπ + b(Xiπ )h + σ(Xiπ )∆i W + [σσ 0 ](Xiπ ) (∆i W )2 − h
2
1 1
+ b0 σ(Xiπ )∆i Z + [bb0 + b00 σ 2 ](Xiπ )h2
2 2
1
+ [bσ 0 + σ 00 σ 2 ](Xiπ ) (∆i W h − ∆i Z)
2
1 2 00 0 2 π 1 2
+ [σ σ + σ(σ ) ](Xi ) (∆i W ) − h ∆i W
2 3
R ti+1
where ∆i Z = ti (Ws − Wti )ds, ∆i W = Wti+1 − Wti .
This is an Ito-Taylor scheme with strong order 1.5 (no proof required!).
Z t Z t
Xt = X0 + b(Xs )ds + σ(Xs )dWs , t≤T,
0 0
representing the price of some stock. We assume a deterministic interest rate r and
1. Recall the price of an european option with payoff g(XT ) at maturity T , where
146
RT
2. We consider now options written on AT = 0 Xs ds. Using a grid π with
X π . Show that
1 C
E[|g(AT ) − g(AπT )|p ] p ≤ √ ,
n
(a) Write down the Euler scheme for the Black-Scholes model. Explain one
drawback of this scheme and why, in fact, we don’t need to use it.
1 C
E[|g(AT ) − g(AπT )|p ] p ≤ .
n
T
We denote by X n its Euler scheme using a grid π with constant time step n. We
consider a bounded measurable function f and assume that the following expansion
holds true:
c1 1
E[f (XT )] = E[f (XTn )] + + O( 2 ) . (6.83)
n n
2. One would like to take advantage of the expansion (6.83). Suggest a new
147
3. One would like to implement this new approximation in practice.
increments. Give the expression of the MSE and the asymptotic variance
ii. Give the expression of the MSE and the asymptotic variance of the
Exercise II.11. (Lookback Option) Let X be the solution of the one dimensional
SDE:
Z t
Xt = X0 + σ(Xs )dWs ,
0
representing the price of some underlying asset. We assume that σ is Lipschitz and
positive. We would like to approximate the price of an option with the following
We assume that
π
E g( max Xt ) ' E g( max Xt ) .
t∈[0,T ] t∈[0,T ]
We know how to simulate Xtπi , ti ∈ π. The goal here is to understand how to simulate
then
148
knowing that (no proof required!)
Law( max Xtπ | (Xtπj )j≤n ) = Law( max Xtπ | Xtπi , Xtπi+1 )
t∈[ti ,ti+1 ] t∈[ti ,ti+1 ]
P( sup Wt ≥ z, Wt ≤ y) = P(Wt ≥ 2z − y) .
t∈[0,T ]
2
P( sup Wt ≥ z, | Wt = y) = e− T z(z−y) .
t∈[0,T ]
3. Explain how to compute E g(maxt∈[0,T ] Xtπ ) in practice.
Exercise II.12. Let b and σ be Cb2 functions from R to R and X be the solution of
Z t Z t
Xt = X0 + b(Xs )ds + σ(Xs )dWs , 0 ≤ t ≤ T .
0 0
1. Recall the definition of ∇X the tangent process for X and the equation (E) it
satisfies .
2. Prove that
" #
p
E sup |∇XT | ≤ Cp .
t≤T
149
Exercise II.13. Let X be given by
Z t Z t
Xtx =x+ b(Xsx )ds + σ(Xsx )dWs ,
0 0
2. Write down the Euler scheme for X and for ∇X, on a discrete time grid π.
Z t Z t
Xtx =x+ b(Xsx )ds + σ(Xsx )dWs ,
0 0
2. Show that
" #
2
E sup |∇Xt − ∇Xti | ≤ C|π|
t∈[ti ,ti+1 ]
3. Show that
" #
E sup |∇Xt − ∇Xtπ |2 ≤ C|π| .
t≤T
150
Exercise II.15. We work in the Black-Scholes setting with
where b and σ are Cb2 . For α > 0 and H be a bounded progressively measurable
process, we introduce
Z t Z t
Xtx (α) =x+ (b(Xsx (α)) − αHs σ(Xsx (α))ds + σ(Xsx (α))dWs
0 0
∂X α
The goal of this exercise is to prove that Ut := ∂α α=0 satisfies
dUt = Ut (b0 (Xt )dt + σ 0 (Xt )dWt ) − Ht σ(Xt )dt and U0 = 0 . (6.84)
1. We define
Xt (α + ) − Xt (α)
∂α Xt (α) = lim
→0
151
using a heuristic argument, show that
Z t
∂α Xtx (α) = b0 (Xsx (α))∂α Xtx (α) − αHs σ 0 (Xsx (α))∂α Xtx (α) − Hs σ(Xsx (α) ds
0
Z t
+ σ 0 (Xsx (α))∂α Xtx (α)dWs
0
Z 1 Z 1
0
b̃t = b (Xt + λ(Xt − Xt ))dλ , σ̃t = σ 0 (Xt + λ(Xt − Xt ))dλ.
0 0
(a) Write down the dynamics of Γ using b̃ , σ̃ and show that, for ∈ [−1, 1]
" #
E sup |Γt |p ≤ Cp .
t≤T
Ut
4. Compute the dynamics of ∇Xt and give the expression of U using (∇Xt ).
Exercise II.17. 1. Prove that the estimator given in Proposition ?? has finite
variance.
Exercise II.19. 1. Prove that any L2 -optimal quantizer is stationary, recall (4.7).
152
Exercise II.20. Prove Proposition 5.1.
153
Part III
τM := inf{t ≥ 0 | |Xtπ | ≥ M } ∧ T ,
XtM,π := Xt∧τ
π
M
.
We observe that
Z t∧τM Z t∧τM
XtM,π = x + b(s̄, Xs̄π )ds + σ(s̄, Xs̄π )dWs ,
0 0
Z t Z t
=x+ b(s̄, Xs̄M,π )1{τM >s} ds + σ(s̄, Xs̄M,π )1{τM >s} dWs .
0 0
p
Using Hölder inequality (recalling that p > 2 ≥ 1),
Z t Z t
M,π p p M,π
M,π p p p−1 −1 p
E sup |Xu | ≤ Cp (x + t E |b(s, Xs̄ )| ds + t 2 E |σ(s, Xs̄ )| ds
0≤u≤t 0 0
154
which a fortiori leads to
Z t
M,π p M,π p
E sup |Xu | ≤ Cp (1 + E sup |Xu | du) .
0≤u≤t 0 0≤u≤s
1. We have
π 1
Xi+1 = Xiπ + b(Xiπ )(ti+1 − ti ) + σ(Xiπ )(Wti+1 − Wti ) + a(Xiπ )((Wti+1 − Wti )2 − (ti+1 − ti )) .
2
and
Z u Z u
T (u) = (b(Xs ) − b(Xs̄ ))ds + (σ(Xs ) − σ(Xs̄ ) − a(Xs̄ )(Ws − Ws̄ ))dWs
0 0
The only new extra term in the analysis, compared with the Euler scheme is
Z t
At := (a(Xs̄ ) − a(Xs̄π ))(Ws − Ws̄ )dWs .
0
155
We then compute, using BDG inequality,
Z t
p
p π 2
E sup |Au | ≤ Cp E ( |a(Xs̄ ) − a(Xs̄ ))(Ws − Ws̄ )| 2 ) ds
0≤u≤t 0
Z t
π p
≤ Cp E |a(Xs̄ ) − a(Xs̄ ))(Ws − Ws̄ )| ds
0
Z t
π p p
≤ Cp E |a(Xs̄ ) − a(Xs̄ )| Es̄[|Ws − Ws̄ | ] ds
0
Z t
≤E |a(Xs̄ ) − a(Xs̄π )|p ds .
0
Since σ is bounded and Cb2 , we have that a is Lipschitz and we finally obtain
E sup |Au |p ≤ Cp E sup |δXu |p .
0≤u≤t u≤t
The proof is then concluded using the same arguments as the ones used for
1.
Z t Z t Z t
2 1 φ(t) B
1
( φ (s)ds) ≤ (φ(t)
2 φ(s)ds) ≤ 2+ φ(s)ds ,
0 0 2B 2 0
where we used Young’s inequality for the last part. Finally, we obtain
Z t
2
φ(t) ≤ (2A + B ) φ(s)ds + 2C .
0
156
1. Apply Ito’s formula to b(Xs ) and σ(Xs ) and discretize the integrals. We sketch
1
Xti+1 ' Xti + σ(Xti )∆i W + σ 00 σ 2 (∆i W h − ∆i Z)
2
Z ti+1 Z t
+ σ 0 σ(Xs )dWs dWt
ti ti
1
Xti+1 ' Xti + σ(Xti )∆i W + σ 00 σ 2 (∆i W h − ∆i Z)
2
Z ti+1 Z t
+ σ 0 σ(Xti ) dWs dWt
ti ti
Z ti+1 Z tZ s
+ (σ 0 σ)0 σ(Xti ) dWr dWs dWt
ti ti ti
2. Gaussian vector s.t. E[∆i Z] = 0, E |∆i Z|2 = 13 ∆3 and E[∆i Z∆i W ] = 12 ∆2 .
√ 3
3. Set ∆i W = hG1 , ∆i Z = 12 h 2 (G1 + √1 G2 ),
3
(G1 , G2 ) ∼ N (0, I2 ).
1. EQ e−T r g(XT ) , Xi+1
π = X π + rh + σ(X π )∆ W Q .
i i i
2. Euler scheme.
157
(b) We first observe that
1 1
E[|g(AT ) − g(AπT )|p ] p ≤ CE[|AT − AπT |p ] p
Z T
AT − AπT = (Xs − Xs̄ )ds
0
We then have
Z T Z T
∗ ∗
E[|AT − AπT |p ] ≤2 p−1
(E | (s − s)rXs ds| p
+E | p
(s − s)σXs dWs | )
0 0
Z T Z T
p
≤ Cp ( |π|p E sup |Xu |p ds + E | |(s∗ − s)σXs | 2 ds|2 )
0 u 0
Z T
≤ Cp |π| E sup |Xu | ds ≤ Cp |π|p .
p p
0 u
1 PN n j
1. The ’classical’ estimator is N j=1 f ((XT ) ) and then
2. We have that
1
E 2f (XT2n ) − f (XTn ) = E[f (XT )] + O( 2 ) .
n
3. (a) We have
158
because of independence. Then we have that, for n big,
(b) We have
1. We compute
2. We observe that
M − Xtπi
max Xtπ ≤ M ⇔ max Wt − Wti ≤
t∈[ti ,ti+1 ] t∈[ti ,ti+1 ] σ(Xtπi )
M − xi xi+1 − xi
P( max Xtπ ≤ M | Xtπi = xi , Xtπi+1 = xi+1 ) = P( max Wtti ≤ | Wtti+1
i
= )
t∈[ti ,ti+1 ] t∈[ti ,ti+1 ] σ(xi ) σ(xi )
M − xi xi+1 − xi
= 1 − P( max Wtti ≥ | Wtti+1
i
= )
t∈[ti ,ti+1 ] σ(xi ) σ(xi )
(M −xi )(M −xi+1 )
−2
hσ 2 (xi )
=1−e
159
using the previous question.
3. We use the inverse transform method. Once the values Xtπi have been simu-
1 π q
Xti + Xtπi+1 + (Xtπi+1 − Xtπi )2 − 2hσ 2 (Xtπi ) ln(U i ))
2
Remark: (see also lecture notes section 2.7, proof of Proposition 2.3) The
above proof should be applied to the localised version of Y i.e. Y·∧τn with
τn = inf{t ≥ 0 | |Yt | ≥ n} ∧ T ,
160
3. Let Y and Z be solution of (6.85) (with b = 0) and ∆ = Y − Z. We then
observe that
Z t
∆t = σ 0 (Xsx )∆s dWs
0
We then have
Z u
sup |∆u |2 = sup | σ 0 (Xsx )∆s dWs |2 ,
u≤t u≤t 0
Z t
2 0 x 2
E sup |∆u | ≤ 4E |σ (Xs )∆s | ds .
u≤t 0
Z t
2 2
E sup |∆u | ≤ C E sup |∆u | ds .
u≤t 0 u≤s
And then using Gronwall’s Lemma, we have that E supu≤t |∆u |2 = 0.
Xtx+ −Xtx
1. It is the derivative of X x with respect to x i.e. ∇Xtx = lim→0 .
3. We compute that
R ti σ 0 (Xs )2 Rt
(b0 (Xs )− )ds+ 0 i σ 0 (Xs )dWs
∇Xtxi = e 0 2
σ 0 (Xtπ )2
Pi−1 0 j Pi−1
dπ
π
j=0 (b (Xt )− )(tj+1 −tj )+ j=0 σ 0 (Xtπ )(Wtj+1 −Wtj )
∇X ti =e j 2 j
161
1. the strong convergence has a rate equal to 1/2.
to get
" # Z
T
E sup |δt | 2
≤ CE |(σ 0 (Xs ) − σ 0 (Xsπ ))∇Xs |2 + |σ 0 (Xsπ )(∇Xs − ∇Xs̄ )|2 ds .
t≤T 0
We then get
" # Z
T
2 0 0
E sup |δt | ≤ CE |(σ (Xs ) − σ (Xsπ ))∇Xs |2 + |∇Xs − ∇Xs̄ )| 2
ds .
t≤T 0
162
and applying CS inequality,
" # Z
1 1 T
2 4 2 0 0 π 4 2 2
E sup |δt | ≤ CE sup |∇Xs | E sup |(σ (Xs ) − σ (Xs ))| +E |∇Xs − ∇Xs̄ )| ds .
t≤T 0
1 log(y/x) − (r − 21 σ 2 )T
h(x, y) = √ ϕ(ζ(x, y)), ζ(x, y) = √
yσ T σ T
And we compute
log(y/x) − (r − 12 σ 2 )T
s(x, y) = ∂x h(x, y)/h(x, y) = −ζ(x, y)∂x ζ(x, y) =
xσ 2 T
Z
∂σ u(σ, x) = g(y)s(σ, y)h(σ, y)dy .
We compute
log(y/x) − r 1 √ √ WT WT
∂σ ζ = − √ + T , ∂σ ζ(XTx ) = T − √ , ζ(XTx ) = √
σ2 T 2 σ T T
1
∂σ h = (− − ζ∂σ ζ)h
σ
and then
x 1 x x
∂σ u = E g(XT )(− − ζ(σ, XT )∂σ ζ(σ, XT ))
σ
x 1 WT
= E g(XT )(− + − WT )
σ σT
163
3. We have
Z Z 2 h(x, y)
2 2 ∂xx
∂xx u = g(y)∂xx h(x, y)dy = g(y) h(x, y)dy
h(x, y)
Recall that
∂x h = −ζ∂x ζh
1. heuristic: assume that ∂α X exists and then differentiate under the integral
2. (say b = 0)
obtain, for p ≥ 2,
Z T
p
E sup |Γu | ≤ Cp E |σ(Xs )|p ds
u≤t 0
Using then classical estimate for the solution of an SDE and the fact that
164
(b) We observe that setting δs = ∂α Xs (0) − Γs ,
Z t Z t
δt = − Hs (σ(Xs ) − σ(Xs ))ds + (σ 0 (Xs )δs + (σ 0 (Xs ) − σ̃s )Γs )dWs +
0 0
Combining the result of 2.a, estimation of |σ 0 (Xs )− σ̃s | (see lecture notes)
And that
165
Similarly one computes
√
pπ0 − pa0 ≤ C π
1.
h i h i
Θ − Θ̂ = E[F (X)|Y ] − E E[F (X)|Y ] |Ŷ + E F (X) − F (X̂)|Ŷ
h i h i
= ϕF (Y ) − E ϕF (Y )|Ŷ + E F (X) − F (X̂)|Ŷ
h i
kϕF (Y ) − E ϕF (Y )|Ŷ k2 ≤ kϕF (Y ) − ϕF (Ŷ )k2
≤ [ϕF ]kY − Ŷ k2
h i
kE F (X) − F (X̂)|Ŷ k2 ≤ kF (X) − F (X̂)k2
≤ [F ]kX − X̂k2
166
Using Doob’s inequality, we get easily that Γ ∈ S 2 as b is bounded. It also clear
Rt
showing that Γt Yt + 0 cr Γr dr is a local martingale, which is - in fact - a martingale
Z t Z T
Γt Yt + cr Γr dr = E ΓT YT + cr Γr dr Ft ,
0 0
and ζ = ξ 0 − ξ, we have
Z T Z T
Ut = ζ + f 0 (r, Yr0 , Zr0 ) − f (r, Yr , Zr ) dr − Vr dWr .
t t
We observe that
f 0 (r, Yr0 , Zr0 ) − f (r, Yr , Zr ) = f 0 (r, Yr0 , Zr0 ) − f 0 (r, Yr , Zr0 ) + f 0 (r, Yr , Zr0 ) − f 0 (r, Yr , Zr )
(i)
For 0 ≤ i ≤ d, we consider the vector Zr whose last d − i components are those of
167
Zr0 and the first i components are those of Zr . For 1 ≤ i ≤ d, we set
(i−1) (i)
f 0 r, Yr , Zr − f 0 r, Yr , Zr
bir = 1{Vri 6=0} .
Vri
Z T Z T
Ut = ζ + (ar Ur + Vr br + cr ) dr − Vr dWr ,
t t
Z T
−1
Ut = Γt E ζΓT + cr Γr dr Ft ,
t
with for 0 ≤ r ≤ T ,
nZ r
1
Z r Z r o
2
Γr = exp bu · dWu − |bu | du + au du .
0 2 0 0
If moreover, U0 = 0 we have
Z T
0 = E ζΓT + cr Γr dr ,
0
and the r.v. is non negative. Then, it is equal to zero P a.s. which implies ζ = 0
Z ti+1
1 2
ζie = − σ 2 (Xtπi ) − σ 2 (Xsπ ) ∂xx u(s, Xsπ )ds (6.86)
2 ti
168
has already been studied in the proof of the weak error for the Euler scheme. The
2 2
σ 2 (Xtπi ) − σ 2 (Xsπ ) ∂xx u(s, Xsπ ) = σ 2 (Xtπi ) − σ 2 (Xsπ ) (∂xx u(s, Xsπ ) − ∂xx
2
u(ti , Xtπi )) =: Ai
2
+ σ 2 (Xtπi ) − σ 2 (Xsπ ) ∂xx u(ti , Xtπi )) =: Bi
We have Eti[|Ai |] ≤ β|π| and applying Ito’s formula, we also obtain |Eti[Bi ] | ≤ β|π|.
h i Z s
(0) 1
Eti u (s, Xsπ ) =u (0)
(ti , Xtπi ) + Eti (0)
{∂t u (t, Xtπ ) + σ(Xtπi )∂xx
2
u(t, Xtπ )}dt
ti 2
also
(0) ∆Wi
Z̃i = Eti u (ti+1 , Xtπi+1 )
ti+1 − ti
∆Wi
Now we compute, setting Hi := hi
Z Z
ti+1
1 σ(Xtπi ) ti+1
Eti u(ti+1 , Xti+1 )Hi = Eti Hi [∂t + σ 2 (Xtπi )∂xx
2
]u(t, Xtπ )dt + ∂x u(t, Xtπ )dt
ti 2 hi ti
(6.88)
169
Observe that
Z ti+1
1 2 π 2 π
|Eti Hi [∂t + σ (Xti )∂xx ]u(t, Xt )dt | (6.89)
ti 2
Z ti+1
1 2 π 2 π 1 2 π 2 π
= |Eti Hi {[∂t + σ (Xti )∂xx ]u(t, Xt ) − [∂t + σ (Xti )∂xx ]u(ti , Xti )}dt |
ti 2 2
(6.90)
≤ C|π| (6.91)
and that
Z t
1 2 π 2
Eti[∂x u(t, Xt )] = Eti ∂x u(ti , Xti ) + [∂t + σ (Xti )∂xx ]∂x u(s, Xs )ds (6.92)
ti 2
We thus get
Z
σ(Xtπi ) ti+1
|Eti ∂x u(t, Xtπ )dt − Vtii | ≤ C|π| (6.93)
hi ti
170