Professional Documents
Culture Documents
section (Introduction) : Diffusions Arise in Various Application Areas. They Are
section (Introduction) : Diffusions Arise in Various Application Areas. They Are
section (Introduction) : Diffusions Arise in Various Application Areas. They Are
\usepackage[utf8]{inputenc}
\usepackage{geometry, amsmath}
\numberwithin{equation}{section}
\usepackage{amsfonts, amssymb}
\usepackage{enumitem}
\usepackage{algorithm}
\usepackage{mathtools}
\usepackage{etoolbox}
\usepackage{pseudocode}
\usepackage[noend]{algpseudocode}
\usepackage[square]{natbib}
\usepackage[section]{placeins}
\usepackage{tcolorbox}
\usepackage[framemethod=TikZ]{mdframed}
\usepackage{lipsum}
\usepackage{theorem}
\usepackage{hyperref}
\mdfdefinestyle{MyFrame}{%
linecolor=blue,
outerlinewidth=2pt,
roundcorner=20pt,
innertopmargin=\baselineskip,
innerbottommargin=\baselineskip,
innerrightmargin=20pt,
innerleftmargin=20pt,
backgroundcolor=gray!50!white}
\title{\textbf{18-Months Upgrade Report}}
\author{Divakar Kumar }
\date{March 2016}
\newtheorem{theorem}{Theorem}[section]
\newtheorem{definition}{Definition}[section]
\newtheorem{lemma}[theorem]{Lemma}
\newtheorem{proposition}[theorem]{Proposition}
\newtheorem{corollary}[theorem]{Corollary}
\newenvironment{proof}[1][Proof]{\begin{trivlist}
\item[\hskip \labelsep {\bfseries #1}]}{\end{trivlist}}
\newenvironment{example}[1][Example]{\begin{trivlist}
\item[\hskip \labelsep {\bfseries #1}]}{\end{trivlist}}
\newenvironment{remark}[1][Remark]{\begin{trivlist}
\item[\hskip \labelsep {\bfseries #1}]}{\end{trivlist}}
\newcommand{\qed}{\blacksquare }
\begin{document}
\maketitle
\section{Introduction}
\section{Introduction}
Diffusionsariseinvariousapplicationareas.Theyare
widelyusedinmyriadfieldsincludingfinance,economics,
physics\citep{Oksendal2003}\citep{Kloeden1999}and
biologicalsciences\citep{Jenkins2010}.Motionofamolecule
\citep{Maki},fluctuationsinthestockmarket\citep{Black1973}are
commoninstancesstudiedthroughthediffusionprocesses.Sometimes
discreteprocessessuchasthepopulationgrowthcanbeverywell
approximatedbydiffusionprocessesasthetimediscretizationtendsto
becomenegligiblysmall\citep{Tuckwell1974}.Thetransitiondensityof
diffusionstypicallyisunknown,therefore,Eulerdiscretizationmethod
iscustomarilyusedforstudyingtheunderlinedbehaviour.Thestudyof
intractablediffusionsmodelusingdiscretizationmethodsofteninvolves
MonteCarlosimulationsinturngivingrisetoMonteCarloand
discretizationerrors.Controllingdiscretizationerroris
computationallyexpensiverenderingtimediscretizationsufficiently
negligible\citep{Kloeden1999}.Thusweseethatapproximationmethods
forsimulationofdiffusionsarenotonlyinexactbutalsoinvolve
enormouscomputationalcost.Consequentlytheneedtopreciselysimulate
trajectoriesofdiffusionswithoutanapproximationerrorarises.Exact
algorithms\citep{Beskos2005,Beskos2006}stipulatemethodtoprecisely
simulatethetrajectoriesofdiffusions.Theyareaclassof
retrospectiveMonteCarlomethodsfordiffusionforsimulatingand
characterizingsamplepathsofdiffusionsoverfinitetimeintervalata
finitecollectionofpoints.Thismethodutilisesrejectionsamplingon
thediffusionpathspace.Whichconsequentlyreliesonfindingsuitable
proposalmeasurestodrawtheproposalsamplepathsaccurately.\\
Considerableinterestinthestudyofintractabledistributionsand
generatingsamplesfromthemwasshown.\citep{Roberts1996}proposeda
diffusionsbasedmethodtoapproximatetheintractabledistributions.
\citep{Roberts1996}considered\textit{Langevindiffusion}whose
invariantdistributionisthedistributionoftheinterest.Precisely
simulatingthetrajectoriesofsuch\textit{Langevindiffusion}isseen
asacircularproblem\citep{Roberts2016}.TheScalableLangevinExact
(ScaLE)\citep{Roberts2016}isanovelalternativetogradientbased
LangevinMCMCschemessuchasMALA\cite{Roberts1996}whichcircumvents
theneedtouseMetropolistypecorrection.Consequentlythe
circumventionofthismethodologyishighlyapplicableto`BigData'
problems.TheScaLEmethodapproximatestheintractabledistributionof
interestwithaquasistationarydistributionofa`killed'
\textit{Langevindiffusion}.\\
Theaimofthisreportistofurnishanintroductiontothecurrent
literatureonexactsimulationfromdiffusionstrajectories.Furthermore
thereportintroducesResampledScalableLangevinExact(ReScaLE)
methodtoapproximateintractabledistributionsofinterestusingrecent
advancesinthestudyofquasistationarybehavior\citep{Blanchet2012}.
The next section walks us through some of the techniques and concepts
utilised during the designing of the algorithms. The first part of the
section 2 explains Rejection sampling used in the construction of the
Exact algorithm. Afterwards it explains the concepts involved in Exact
algorithm. Instances include Brownian motion, Brownian Bridge, and
Poisson processes. Section 3 attempts at explaining the methodology used
in the Exact algorithm, with respect to the application extended to the
diffusion process with drift function $1/(1+x^2)$. Section 4 discusses
the working of ScaLE algorithm combined with its implementation towards
the case of the Cauchy distribution.
\begin{definition}
A It\^{o} diffusion $\{X_t \, | t \geq 0\}$ is characterized by the SDE
\begin{align}
dX_t = \mu(X_t, t) dt + \sigma(X_t, t) dB_t
\end{align}
is a stochastic process defined on $(\mathbf{\Omega}, \mathcal{F},
\mathbb{P})$ such that
\begin{itemize}
\item $\mu(X_t, t)$ is $\mathcal{F}_t$ measurable and
$\mathbb{P}\left( \int \limits_0^t |\mu(X_s, s)| ds < \infty, \forall \,
t \geq 0\right) = 1$
\item $\mathbb{P}\left(\int \limits_0^t |\sigma(X_s, s)|^2 ds <
\infty, \forall \, t \geq 0\right) = 1$
\end{itemize}
\end{definition}
Next we discuss an important result that best describes the changes in
the behavior of a It\^{o} diffusion defined on $(\mathbf{\Omega},
\mathcal{F}, \mathbb{P})$ if there is a change in the probability
measure from $\mathbb{P}$ to $\mathbb{Q}$. We can informally apply this
transformation to a given It\^{o} diffusion to obtain another. It is
\^{o} diffusion facilitating simulation studies. These transformations
keep solutions unchanged but transformed SDE will have different drift
and diffusion coefficient since the underlined probability measure is
different.
\begin{theorem}[Girsanov's Theorem]
Consider a It\^{o} diffusion $dX_t = \mu(X_t, t) dt + \sigma(X_t, t)
dB_t$ defined on $(\mathbf{\Omega}, \mathcal{F}, \mathbb{P})$. A driftless It\^{o} diffusion $Y_t = \sigma(Y_t, t) dB_t$ on probability space
$(\mathbf{\Omega}, \mathcal{F}, \mathbb{Q})$ satisfies,
\begin{align}
\frac{d\mathbb{P}}{d\mathbb{Q}}(X_t) = \exp\left( \int \limits_0^t
\frac{\mu(X_t, t)}{\sigma^2(X_t, t)}dX_t - \frac{1}{2}\int
\limits_0^t \frac{\mu^2(X_t, t)}{\sigma^2(X_t, t)}dt \right)
\end{align}
\end{theorem}
Now we discuss about Poisson process which is a counting process on $[0,
\infty)$. A Poisson process counts the number of times an event has
occurred in a given time interval. It is one of the most prominent
stochastic process used to model arrival process of customers or calls
in a telephone exchange \citep{Ross2010}.
\begin{definition}
A Poisson process of varying arrival rate $\lambda(t)$ is a continuous
time stochastic process $\{N(t): t\geq 0 \}$ which counts the number of
events in $[0,t]$ and it satisfies the following properties:
\begin{enumerate}
\item[\textbf{P1}. ] $\{N(t): t\geq 0 \}$ has independent increment
property, i.e., for any partition $P_n:= \{0=t_0<t_1<...<t_n=t\}$ of the
interval $[0,t]$ for some $t>0$, the random variables of counts $\{{N}
(t_{i-1}, t_{i}):= N(t_i)-N(t_{i-1})|i = 1,2,...,n \}$ are mutually
independent.
\limits_{0}^{1} f(x, y) \, dy
\end{align}
$f(x)$ in \eqref{eq:2.1.1} can be viewed as the marginal density of
$f(x,y) = \mathbb{I}_{\{0 \leq y \leq f(x)\}}$. Thus $f(x)$ is the
marginal density of uniform distribution on rectangle $\{(x,y): 0 \leq y
\leq f(x)\}$ given by the area under the graph. We can generate samples
from $f$ by sampling within the area under the curve. Above idea can not
be translated to unbounded densities or densities with unbounded
support. Thus we choose a proposal density $g$ in such a way that
$\exists M >0$ with
\begin{align} \label{eq:2.1.2}
f(x) \leq g(x) M \quad \forall \, x
\end{align}
Thus instead of sampling directly from $f$, we sample $X$ from $g$ and
accept the sampled values with probability $P = \frac{1}{M}\frac{f(X)}
{g(X)}$ otherwise reject $X$ as the candidate sample from $f$. Thus
simplest version of rejection sampling is
\begin{pseudocode}[Ovalbox]{\textbf{Rejection Sampling Method}}{M, f, g}
1. \quad X \sim g \\
2. \quad u \sim U[0,1] \\
3. \quad \IF u \leq \frac{1}{M}\frac{f(X)}{g(X)} \quad \GOTO 4. \quad
\mbox{otherwise} \quad \GOTO 1. \\
4. \quad \RETURN{X}
\end{pseudocode}
\begin{proof}
Let $\Omega$ denote the set of all possible values of $X$. Assume $A
\subset \Omega$. Probability that values of $X$ lies in $A$ given that
it was accepted is shown by
\begin{align} \label{eq:2.1.3}
P\left(X \in A \left| \right. u \leq \frac{1}{M}\frac{f(X)}{g(X)}\right)
&= \frac{P\left(X \in A, u \leq \frac{1}{M}\frac{f(X)}{g(X)}\right)}
{P\left(u \leq \frac{1}{M}\frac{f(X)}{g(X)}\right)} \\ \label{eq:2.1.4}
&= \frac{\int \limits_{A} {\frac{1}{M}\frac{f(y)}{g(y)}} g(y) dy}
{\int \limits_{\Omega} {\frac{1}{M}\frac{f(y)}{g(y)}} g(y)
dy}\\ \label{eq:2.1.5}
&= \frac{\int \limits_{A} f(y) dy}{\int \limits_{\Omega}^\infty f(y) dy}
= P\left(X \in A\right)
\end{align}
The last equality \eqref{eq:2.1.5} follows since denominator in
\eqref{eq:2.1.5} equals one (We are integrating over all possible values
of X ). Thus simulating samples $X_i \sim g$ and retaining the sample if
$u_i \leq \frac{1}{M}\frac{f(X_i)}{g(X_i)}$ gives i.i.d sample from
target $f$.
\end{proof}
\begin{figure}[h]
\centering
\includegraphics[scale=0.5]{Rplot.png}
\caption{An illustration of rejection sampling}
\end{figure}
Figure 1 gives an intuitive explanation of rejection sampling. Suppose
we want to uniformly simulate points from the subgraph. It is denoted by
yellow region in the graph but due to the difficult simulation from the
density $f$, we simulate points uniformly from the graph with yellow and
grey region together. Simulating points uniformly from the subgraph $\
{(x,u)\in S_q \times R^+:q(x)\leq u\}$ is achieved by first simulating
points $X_1, X_2, ...$ from the support of $q$ i.e. $S_q$ and then
drawing $u_1, u_2, ...$ independently according to $U[0, Mq(X_i)]$ for
$i \in \{1,2,...\}$. This gives coordinates $\{(X_i,u_i): i \in \
{1,2,...\}\}$ on the graph. Next we retain points according to their
position on the subgraph. If point $(X_i,u_i)$ fall in the yellow area
then $X_i$ is retained as candidate sample from target density $p$. Each
simulated point $X_i \sim q$ has acceptance probability $\frac{1}
{M}\frac{p(X_i)}{q(X_i)}$ of being accepted as candidate sample from the
target density $p$. Algorithm stops until $N$ points has been retained.
This method gives i.i.d sample from target density which can be
illustrated as follow: