Professional Documents
Culture Documents
Full Ebook of Foundations of Quantitative Finance Book Ii Probability Spaces and Random Variables 1St Edition Robert R Reitnao Online PDF All Chapter
Full Ebook of Foundations of Quantitative Finance Book Ii Probability Spaces and Random Variables 1St Edition Robert R Reitnao Online PDF All Chapter
Full Ebook of Foundations of Quantitative Finance Book Ii Probability Spaces and Random Variables 1St Edition Robert R Reitnao Online PDF All Chapter
https://ebookmeta.com/product/foundations-of-quantitative-
finance-book-i-measure-spaces-and-measurable-functions-1st-
edition-robert-r-reitano/
https://ebookmeta.com/product/probability-random-variables-and-
random-processes-4th-edition-hwei-p-hsu/
https://ebookmeta.com/product/foundations-of-quantitative-
finance-book-iii-the-integrals-of-riemann-lebesgue-and-riemann-
stieltjes-1st-edition-robert-r-reitano/
https://ebookmeta.com/product/schaum-s-outline-of-probability-
random-variables-and-random-processes-4th-edition-hwei-p-hsu/
Probability Random Variables and Data Analytics with
Engineering Applications P. Mohana Shankar
https://ebookmeta.com/product/probability-random-variables-and-
data-analytics-with-engineering-applications-p-mohana-shankar/
https://ebookmeta.com/product/probability-random-variables-and-
data-analytics-with-engineering-applications-p-mohana-shankar-2/
https://ebookmeta.com/product/probability-and-random-
processes-4th-edition-geoffrey-r-grimmett/
https://ebookmeta.com/product/random-matrices-and-non-
commutative-probability-1st-edition-arup-bose/
https://ebookmeta.com/product/learning-modern-c-for-finance-
foundations-for-quantitative-programming-fourth-early-
release-4th-edition-daniel-hanson/
Foundations of Quantitative
Finance
Chapman & Hall/CRC Financial Mathematics
Series
Aims and scope:
The field of financial mathematics forms an ever-expanding slice of the financial sector. This series aims
to capture new developments and summarize what is known over the whole spectrum of this field. It
will include a broad range of textbooks, reference works and handbooks that are meant to appeal to both
academics and practitioners. The inclusion of numerical code and concrete real-world examples is highly
encouraged.
Robert R. Reitano
Brandeis International Business School
Waltham, MA
First edition published 2023
by CRC Press
6000 Broken Sound Parkway NW, Suite 300, Boca Raton, FL 33487-2742
Reasonable efforts have been made to publish reliable data and information, but the author and publisher cannot as-
sume responsibility for the validity of all materials or the consequences of their use. The authors and publishers have
attempted to trace the copyright holders of all material reproduced in this publication and apologize to copyright holders
if permission to publish in this form has not been obtained. If any copyright material has not been acknowledged please
write and let us know so we may rectify in any future reprint.
Except as permitted under U.S. Copyright Law, no part of this book may be reprinted, reproduced, transmitted, or
utilized in any form by any electronic, mechanical, or other means, now known or hereafter invented, including pho-
tocopying, microfilming, and recording, or in any information storage or retrieval system, without written permission
from the publishers.
For permission to photocopy or use material electronically from this work, access www.copyright.com or contact the
Copyright Clearance Center, Inc. (CCC), 222 Rosewood Drive, Danvers, MA 01923, 978-750-8400. For works that are
not available on CCC please contact mpkbookspermissions@tandf.co.uk
Trademark notice: Product or corporate names may be trademarks or registered trademarks and are used only for
identification and explanation without intent to infringe.
DOI: 10.1201/9781003260547
Typeset in CMR10
by KnowledgeWorks Global Ltd.
Publisher’s note: This book has been prepared from camera-ready copy provided by the authors.
to Dorothy and Domenic
Taylor & Francis
Taylor & Francis Group
http://taylorandfrancis.com
Contents
Preface xi
Author xiii
Introduction xv
1 Probability Spaces 1
1.1 Probability Theory: A Very Brief History . . . . . . . . . . . . . . . . . . . 1
1.2 A Finite Measure Space with a “Story” . . . . . . . . . . . . . . . . . . . . 2
1.2.1 Bond Loss Example . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.3 Some Probability Measures on R . . . . . . . . . . . . . . . . . . . . . . . . 11
1.3.1 Measures from Discrete Probability Theory . . . . . . . . . . . . . . 11
1.3.2 Measures from Continuous Probability Theory . . . . . . . . . . . . 16
1.3.3 More General Probability Measures on R . . . . . . . . . . . . . . . 20
1.4 Independent Events . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
1.4.1 Independent Classes and Associated Sigma Algebras . . . . . . . . . 24
1.5 Conditional Probability Measures . . . . . . . . . . . . . . . . . . . . . . . 27
1.5.1 Law of Total Probability . . . . . . . . . . . . . . . . . . . . . . . . . 28
1.5.2 Bayes’ Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
vii
viii Contents
References 249
Index 253
Taylor & Francis
Taylor & Francis Group
http://taylorandfrancis.com
Preface
The idea for a reference book on the mathematical foundations of quantitative finance
has been with me throughout my professional and academic careers in this field, but the
commitment to finally write it didn’t materialize until completing my first “introductory”
book in 2010.
My original academic studies were in “pure” mathematics in a subfield of mathemati-
cal analysis, and neither applications generally nor finance in particular were even on my
mind. But on completion of my degree, I decided to temporarily investigate a career in ap-
plied math, becoming an actuary, and in short order became enamored with mathematical
applications in finance.
One of my first inquiries was into better understanding yield curve risk management, ulti-
mately introducing the notion of partial durations and related immunization strategies. This
experience led me to recognize the power of greater precision in the mathematical specifica-
tion and solution of even an age-old problem. From there my commitment to mathematical
finance was complete, and my temporary investigation into this field became permanent.
In my personal studies, I found that there were a great many books in finance that
focused on markets, instruments, models and strategies, and which typically provided an
informal acknowledgment of the background mathematics. There were also many books
in mathematical finance focusing on more advanced mathematical models and methods,
and typically written at a level of mathematical sophistication requiring a reader to have
significant formal training and the time and motivation to derive omitted details.
The challenge of acquiring expertise is compounded by the fact that the field of quanti-
tative finance utilizes advanced mathematical theories and models from a number of fields.
While there are many good references on any of these topics, most are again written at
a level beyond many students, practitioners and even researchers of quantitative finance.
Such books develop materials with an eye to comprehensiveness in the given subject matter,
rather than with an eye toward efficiently curating and developing the theories needed for
applications in quantitative finance.
Thus the overriding goal I have for this collection of books is to provide a complete and
detailed development of the many foundational mathematical theories and results one finds
referenced in popular resources in finance and quantitative finance. The included topics
have been curated from a vast mathematics and finance literature for the express purpose
of supporting applications in quantitative finance.
I originally budgeted 700 pages per book, in two volumes. It soon became obvious
this was too limiting, and two volumes ultimately turned into ten. In the end, each book
was dedicated to a specific area of mathematics or probability theory, with a variety of
applications to finance that are relevant to the needs of financial mathematicians.
My target readers are students, practitioners and researchers in finance who are quantita-
tively literate, and recognize the need for the materials and formal developments presented.
My hope is that the approach taken in these books will motivate readers to navigate these
details and master these materials.
Most importantly for a reference work, all ten volumes are extensively self-referenced.
The reader can enter the collection at any point of interest, and then using the references
xi
xii Preface
cited, work backwards to prior books to fill in needed details. This approach also works for
a course on a given volume’s subject matter, with earlier books used for reference, and for
both course-based and self-study approaches to sequential studies.
The reader will find that the developments herein are at a much greater level of detail
than most advanced quantitative finance books. Such developments are of necessity typically
longer, more meticulously reasoned, and therefore can be more demanding on the reader.
Thus before committing to a detailed line-by-line study of a given result, it is always more
efficient to first scan the derivation once or twice to better understand the overall logic flow.
I hope the additional details presented will support your journey to better understanding.
I am grateful for the support of my family: Lisa, Michael, David, and Jeffrey, as well as
the support of friends and colleagues at Brandeis International Business School.
Robert R. Reitano
Brandeis International Business School
Author
xiii
Taylor & Francis
Taylor & Francis Group
http://taylorandfrancis.com
Introduction
xv
xvi Introduction
variables and random vectors and the associated distribution functions including, in the
vector case, the variously defined marginal and conditional distribution functions.
The “left continuous” inverse of a distribution function is studied, and will be seen to
be critical in generating samples of random variables in Chapter 4, in the development for
copulas in Chapter 7, and in the study of extreme value theory in Chapter 9. Also included
is a study of independent random variables and sigma algebras, and associated results.
Applying the infinite dimensional product measure space construction of Chapter 9 of
Book I, Chapter 4 develops theoretical frameworks for collections of independent, identically
distributed (i.i.d.) random variables using three constructions. The first construction imple-
ments the Book I approach directly, using the Borel measure induced by the distribution
function of the given random variable X.
After developing ideas underlying stochastic simulations, the second construction for
i.i.d. random variables obtains an infinite dimensional space using the Borel measure induced
by the uniform distribution. The associated independent random variables distributed as
X are then defined with the left continuous inverse of the distribution function of X. The
third construction is a modification of the first in the case where X is discrete, and thus
the original real probability space can be greatly simplified. Both the second and third
constructions modify the Book I development by reducing the original space with essential
subsets.
Chapter 5 then turns to limit theorems for random variable sequences, a topic that will
be extended later in this book, as well as in other books as more tools are introduced.
Initially focusing on binomial random variable sequences, the associated weak law of large
numbers, known as Bernoulli’s theorem, and the strong law of large numbers, known as
Borel’s theorem, are derived.
The chapter then initiates the more general study of modes of convergence of random
variables, investigating convergence in probability, convergence with probability 1, and con-
vergence in distribution. The traditional version of Slutsky’s theorem, to be generalized in
Book VI, and Kolmogorov’s zero-one law on tail events, are derived.
The focus of Chapter 6 is to investigate and rationalize the many notions of “distribution
function.” In Book I, such functions are seen to be characterized by certain properties, and
are both induced by Borel measures and can be used to create Borel measures. In earlier
chapters of the current book, distribution functions are defined by random variables and
can also induce random variables.
Starting with distribution functions on R, this chapter rationalizes these perspectives,
connecting functions with given properties, functions induced by random variables, and
functions induced by Borel measures, in various ways. This investigation is then generalized
to distribution functions on Rn .
Chapter 7 then develops the theory of copulas, which are “simply” joint distribution
functions with uniform marginal distribution functions. This theory has its mathematical
roots in Book I’s Chapter 8 on general Borel measures on Rn . The goal of the chapter is
the derivation of Sklar’s theorem. It states that given any joint distribution function, there
exists a copula, which when valued on that distribution’s marginal distributions, reproduces
the original joint distribution.
Put another way, Sklar’s result states that given only the marginal distributions, all the
complexity of the original joint distribution can be reproduced with the right copula. Since
marginal distributions are relatively easy to estimate, it is no surprise that this theory has
found many applications in finance and elsewhere. Many examples of copulas are illustrated,
and the theory of extreme value copulas is introduced. These results are then extended to
results on tail dependence, where survival copulas are introduced to simplify the analysis.
Critical to the final chapter’s study of extreme value theory, Chapter 8 introduces and
develops some fundamental results on weak convergence of distribution functions, and more
Introduction xvii
Notation 0.1 (Referencing within FQF Series) To simplify the referencing of results
from other books in this series, we will use the following convention.
A reference such as “Proposition I.3.33” is a reference to Proposition 3.33 of Book I,
while Chapter III.4 is a reference to Chapter 4 of Book III, and so forth.
Taylor & Francis
Taylor & Francis Group
http://taylorandfrancis.com
1
Probability Spaces
1
2 Foundations of Quantitative Finance
is intuitively appealing but requires a more formal statement and proof which addresses both
the calculation of flip sequence probabilities, and the meaning of convergence. Certainly not
all such infinite sequences will have this property, and indeed there will be one with all Hs,
and some with 90% Hs, etc.
Notable contributors in this evolution were Jacob Bernoulli (1655–1705) with his Ars
Conjectandi (”The Art of Conjecturing”), published posthumously in 1713; and Abraham
de Moivre (1667–1754) with The Doctrine of Chances in 1718. Bernoulli proved an early
version of the law of large numbers, while De Moivre derived a special case of the central
limit theorem, applying it to approximating binomial probabilities. In other work he also
applied this probability framework to the development of actuarial tables which could be
used to estimate the costs of providing life insurance and other life-contingent benefits.
Another key contributor was Pierre-Simon Laplace (1749–1827) with Théorie analy-
tique des probabilités (”Analytical Theory of Probability”) in 1812, introducing the moment
generating function and the method of least squares, and applying calculus to probability
theory.
The next major break-through required a measure theory, the study of which began with
the work of Henri Lebesgue (1875–1941) and the development of the Lebesgue approach
to integration. This approach was introduced in Book I and will be further developed in
Book III. Lebesgue’s work followed the formal development by Bernhard Riemann (1826–
1866) of the Riemann approach to integration. Using his newly developed theory, Lebesgue
proved that a bounded function f (x) is Riemann integrable on an interval [a, b] if and only if
it is continuous “almost everywhere,” which is to say, except on a set of Lebesgue measure
0. Also critical for such investigations was the formalization of the notion of ”limit” by
Augustin-Louis Cauchy (1789–1857).
The convergence of probability theory and measure theory occurred first in the work of
Andrey Kolmogorov (1903–1987) in his Foundations of the Theory of Probability in 1933.
This work introduced an axiomatic framework for probability theory that both merged and
generalized the discrete and continuous models of the theory. Kolmogorov’s approach is
central to all modern developments of this theory.
µ : E →[0, 1],
1. ∅, S ∈ E;
Probability Spaces 3
2. If A ∈ E, then A
e ∈ E, where A
e denotes the complement of A:
e ≡ {s ∈ S|s ∈
A / A};
S
3. If Aj ∈ E for j = 1, 2, 3..., then j Aj ∈ E.
4. µ(S) = 1;
Remark 1.2
By De Morgan’s laws of Exercise I.2.2, and named for Augustus De Morgan (1806–
1871), it is then also
T the case that the intersection of finitely or countably many events is
an event. That is, j Aj ∈ E since:
\ ^
[
Aj = A
ej .
j j
Remark 1.3 (On probability spaces) In “discrete” sample spaces, which contain
finitely or countably many sample points, E almost always contains each of the sample
points and hence all subsets of S. Consequently discrete probability spaces are always com-
plete. In a general sample space that is uncountably infinite, the collection of events will
virtually always be a proper subset of the collection of all subsets in the same way that the
sigma algebra ML of Lebesgue measurable sets on R is a proper subset of the power set
σ(P (R)). See Section I.2.4.
Given the generality of the sigma algebra structure of the collection of events, it is clear
that on any given sample space, any number of sigma algebras can be defined. The same
statement applies to probability measures. For example, there might be two sigma algebras
E and E 0 with E ⊂ E 0 , so every event in E is an event in E 0 . In that sense, E 0 is a “finer”
sigma algebra because it contains more events, and E a “coarser” sigma algebra because
it contains fewer events. It can also be the case that there are two sigma algebras where
neither E ⊂ E 0 nor E 0 ⊂ E is true.
Some questions that arise in probability spaces include the following, where we note the
answers already developed from Book I.
However, before defining something as the “intersection of all sigma algebras which
possess a given property,” it is critical that it be verified that this collection of sigma
algebras is not empty. If this collection is empty this would render this construction
meaningless. But in all cases the collection of sigma algebras is always non-empty
since it includes the power sigma algebra, and so the existence of a smallest sigma
algebra E implied by this construction is always valid.
same was true for the Borel measure spaces MµF (R) of Chapter I.5 as well as all
constructions under the extension theorems of Chapter I.6 which utilize an outer
measure approach.
However, the Borel sigma algebra B(R) is not complete under either Lebesgue or
Borel measures. The completion theorem of Proposition I.6.20 allows one to always
assume that a sigma algebra is complete, or has been replaced by its “completion,”
knowing that the measures of the original sets will not be changed in the process.
The significance of ensuring that a sigma algebra is complete is that one does not then
need to worry about the effect of sets of measure 0 on the measurability of functions.
As proved in Proposition I.3.16 (by way of Exercise I.3.15), if f : (S, E, µ) → R is a
measurable function with E complete, and g(x) = f (x) except on a set of measure 0,
then g(x) is also measurable.
Example 1.4 (Product sample space) Given a probability space (S, E, µ), define the n-
trial sample space, denoted S n , by:
Intuitively, each point of S n represents an “n-sample” of points from the sample space S,
though formally this is an example of the construction in Chapter I.7. There we constructed
a complete sigma algebra of events E n and probability measure µn ,Yas follows.
n
n 0 n
For E we began with a collection of sets A in S of the form Aj , where Aj ∈ E,
j=1
and on such sets defined a set function µ0 by:
Yn Yn
µ0 Aj = µ(Aj ). (1)
j=1 j=1
This collection of sets is a semi-algebra by Proposition I.7.2, and applying the extension
process of Chapter I.6, the measure space (S n , E n , µn ) is finally constructed in Proposition
I.7.20. The punchline is that A0 ⊂ E n , and on A0 :
µn = µ0 . (2)
This identity is also valid on the algebra A generated by A0 .
Now define a collection of sets {Bk }nk=1 ⊂ A0 by:
Yn
Bk = Ajk ,
j=1
where Ajk ∈ E for all j, k, and Ajk = S when j 6= k. This is a collection of independent
events by Definition 1.15. This means as in (1.26) that for every subset K ⊂ {1, 2, ..., n}
with K ≡ {k1 , k2 , ...km }:
\m Ym
µn Bki = µn (Bki ) . (3)
i=1 i=1
where A0k = Akk for k ∈ K, and A0k = S otherwise. Thus by (1) and (2):
\m Ym Ym
µn Bki = µ(Aki ki ) = µn (Bki ) .
i=1 i=1 i=1
Now let X be a random variable on (S, E, µ), which will simply mean that X : S → R is
measurable in the sense that X −1 (A) ∈ E for every Borel A ∈ B (R) . Then random variables
Probability Spaces 7
{Xj }nj=1 can be defined on (S n , E n , µn ), each with the same distribution function as X, and
such that these random variables are independent by Definition 3.47.
All of the above comments will remain true when n = ∞, by replacing the construction
of Chapter I.7 with that of Chapter I.9.
In Chapter 4, we formalize the notion of generating “random” samples of a random
variable.
P ≡ {B1 , B2 , ..., BN },
denote a bond or loan portfolio where each bond Bj has a loan amount fj , and a risk class
which might be defined in terms of third-party credit ratings for bonds, or internal risk
assessment criteria for other types of loans. For simplicity we refer to these as bonds, and
for each, denote the probability of default in one year by q(Bj ). The probability of receiving
scheduled repayments during the year is then 1 − q(Bj ), and sometimes denoted p(Bj ). Our
goal is to produce a probability space of all possible outcomes of portfolio defaults in one
year.
1. (A First Attempt and Failure) We begin with a natural model of defining the
sample space S = P as the collection of bonds, and the sigma algebra E as the power
sigma algebra E = σ(P (S)). That is, E contains every subset of P and the empty
set ∅. This choice reflects the observation that any subset of S has the potential of
defaulting, including the empty set, and hence every subset is potentially an “event.”
The probability measure µ is defined on E for any subset of S as the probability of
that subset defaulting in one year, and all other bonds making scheduled payments.
So if J ⊂ {1, 2, ..., N } is a subset and the event A is given as:
A = {Bj |j ∈ J},
define: Y Y
µ(A) ≡ q(Bj ) (1 − q(Bj )). (1)
j∈J j∈J 0
Here J 0 denotes all indexes not contained in J. Here and below we explicitly allow
J = ∅.
The reader may note that the formula for µ reflects the assumption that bond defaults
are “independent.” We will develop this notion more formally later, but for now
intuition suffices. If we had N coins, each with probability of heads q(Bj ), then µ(A)
is precisely the probability of obtaining heads only on the jth flips where j ∈ J.
While this formula assigns a probability to each event in E, the resulting structure
(S, E, µ) is not a probability space. The problem with this construction is not E, which
is obviously a sigma algebra, but with µ, which is easily seen to not be a probability
measure on E:
QN
(a) If J = ∅, then A = ∅ but µ(A) = j=1 (1 − q(Bj )) 6= 0 unless at least one
q(Bj ) = 1. Of course we can attempt to fix this by simply declaring the
8 Foundations of Quantitative Finance
3. (A Second Attempt and Success) Given the bond portfolio P, define the sample
space S as the collection of all outcomes of default in one year, specified as N -tuples
of bonds defined by all J and J 0 . Specifically,
Then µ(∅) = 0 by this definition since the empty set produces a vacuous summation.
Further: X
µ(S) = µ(A) = 1,
A∈S
since this is equivalent to:
X hY Y i
q(Bj ) 0
(1 − q(B j )) = 1,
J j∈J j∈J
Probability Spaces 9
It is then also the case that for A ∈ E defined this way, Ae is the collection of default
outcomes not in A, and thus µ(A) = 1 − µ(A) follows from µ(S) = 1. That µ is
e
finitely additive over disjoint elements of E follows by the same argument.
With this construction, (S, E, µ) is now a probability space. Each event in E represents
a subset of the various default outcomes that can occur in one year, while µ(A) for
A ∈ E is the probability that one of these default outcomes is realized in the next
year.
Remark 1.6 (On Example 1.5) It can be argued that a lot of work went into defining
a legitimate probability space in part 3, but that having done this, we did not gain any
additional insights into the default probability model that was contemplated in part 1. Even
without the formal measure-theoretical framework, S most finance practitioners would have
little problem specifying probabilities of events A = i Ai for disjoint {Ai } ⊂ S defined in
part 3, even without the formality offered there.
If there is a lesson in the above example, it is in realizing that a given finance or other
model may well be framed in the units of part 1, whereby we view a bond portfolio as the
sample space and the possible default outcomes in one year as the sigma algebra. While
workable intuitively, and also rigorously in defining probabilities, this framework is techni-
cally not equivalent to a probability space model. What was needed above was a transition
from this “natural” model, to the “probability space” model of a collection of possible default
outcomes as sample points with sets of outcomes as the sigma algebra.
In practice, the details of this construction are sometimes avoided and the basic model
and probabilities specified as in part 1. However, with a little more effort it will always be true
that there exists a probability space (S, E, µ) that is consistent with the specified probabilities.
Example 1.7 (Probability space of bond default losses) As above, let P ≡ {B1 , B2 ,
..., BN } denote a bond or loan portfolio where each bond Bj has a loan amount fj and
probability of default q(Bj ) in one year. In this continuation of the above example we seek
to recognize the various levels of dollar loss upon default, and so the goal is to produce a
probability space which characterizes all possible loss outcomes due to portfolio defaults in
one year.
The loss recognized by the lender on default is often called the loss given default (LGD)
or loss ratio for each loan, denoted lj , and specifies the proportion of the loan amount fj
lost by the lender given a default on loan Bj . Sometimes the loan recovery rate rj is
specified instead, representing the relative amount of fj recovered by the lender given a
default of loan Bj . In general, loss plus recovery need not equal 1.0, meaning lj + rj = 1.0.
This is because the lender will accrue interest for a period after default, and therefore
the amount owed will be (1 + i)fj . For example, if fj = 100 and i = 0.02 and the lender
recovers 60, the recovery rate is rj = 0.6, but the loss ratio is lj = 0.42.
If it is hypothesized that loan Bj has a loss ratio of lj on default, and that this LGD
variable is constant, then the above probability space in part 3 is adequate. We simply define
S 0 ={(LJ , LJ 0 )|J ⊂ {1, 2, ..., N }}, where LJ = {lj fj }j∈J , LJ 0 = {0}j∈J 0 denote the set of
fixed dollar loss outcomes associated with a given default outcomes. The sigma algebra E 0
and µ are then defined as above.
If the loss ratio is not constant for any given bond Bj , then the sigma algebra defined by
default outcomes is too course to identify all possible events. For example, the event defined
10 Foundations of Quantitative Finance
by J ≡ {1} in the constant LGD space is effectively the event that bond B1 defaults. When
LGD is not constant, this “event” is the union of events defined by the different values that
LGD can achieve. Thus what is needed is a refinement of S and E which will in effect split
each of the default events into finer events which identify the loss levels potentially incurred
by the defaulted bonds which each event identifies.
As a simple example of a somewhat more general model, assume that each lj can assume
3 values: lj = 0.25, 0.50, 1.00, on default of Bj . Define a new sample space S 0 in terms
of the original default sample space S of part 3 above. Specifically, if (BJ , BJ 0 ) ∈ S with
index set J = (j1 , j2 , ..., jM ), this sample point will become 3M sample points in S 0 , each one
representing one of the possible loss outcomes for these M bonds. In detail, the 3M points
in S 0 will be of the form:
k1 k2 kM
(LK
J , LJ 0 ) ≡ (lj1 fj1 , lj2 fj2 , ..., ljM fjM , 0, ...0).
Here K ≡ KJ = (k1 , k2 , ..., kM ) denotes one of the 3M M -tuples of points where each ki
equals 1, 2, 3, with respective ljkii -values equal to 0.25, 0.50, 1.00.
So while the original sample space S contained 2N points, equaling the number of subsets
of the collection of N bonds, S 0 will contain 4N points. To see this, note that there are Nj
sample points of S with j defaulted bonds (see (1.9) below), and hence there are Nj 3j
This generalizes, in that if each lj can assume M values, S 0 will contain (M + 1)N points.
Defining E 0 ≡ σ(P (S 0 )), the probability measure on E must also be refined. If A ∈ S 0
contains the bonds defined by a non-empty index subset J = (j1 , j2 , ..., jM ), and loss ratios
specified by KJ = (k1 , k2 , ..., kM ), then define:
Y Y
µ0 (A) = q(Bj ) Pr(kj ) 0
(1 − q(Bj )). (1)
j∈J j∈J
Here given a default of bond Bj , Pr(kj ) denotes the probability that the loss ratio is given
by kj as above.
Now extend µ0 to E 0 additively as in part 3 of Example 1.5.
To prove that:
µ0 (S 0 ) = 1,
note that by (1) this is equivalent to:
X hX Y Y i
q(Bj ) Pr(kj ) (1 − q(Bj )) = 1. (2)
J KJ j∈J j∈J 0
Letting pj,a ≡ Pr(kj = a), and so pj,1 + pj,2 + pj,3 = 1, we have as in Example 1.5:
X hY Y i
1 = q(Bj ) (1 − q(B j ))
J j∈J j∈J 0
X hY Y i
= q(Bj ) (pj,1 + pj,2 + pj,3 ) 0
(1 − q(Bj )) .
J j∈J j∈J
Fixing J: Y X Y
q(Bj ) (pj,1 + pj,2 + pj,3 ) = q(Bj ) Pr(kj ),
j∈J KJ j∈J
and thus:
pi , x = xi ,
f (x) ≡ (1.3)
0, otherwise.
This function has an associated distribution function (d.f.) or sometimes cumula-
tive distribution function (c.d.f.), denoted F (x), and defined:
X
F (x) = f (y), (1.4)
y≤x
To this end, recall that the Chapter I.5 extension process began with defining a set func-
tion |·|F on the semi-algebra A0 of right semi-closed intervals by:
and so:
µ = |·|F , on A0 . (2)
The set function |·|F extends by addition to the algebra A of finite disjoint unions of such
intervals, and is a measure µA on A by Proposition I.5.13, and so by (2):
µ = µA , on A.
Let µF denote the Proposition I.5.23 extension of the measure µA on A to a measure
on MµF (R). Since µ is also well defined on this sigma algebra, Proposition I.6.14 assures
that µ = µF on σ (A) ⊂ MµF (R), the smallest sigma algebra that contains A. Thus µ =
µF on the Borel sigma algebra B(R) since σ (A) = B(R) by Proposition I.8.1. Finally by
Proposition I.6.24, µ = µF on σ CµF (A) , the Proposition I.6.20 completion of σ (A) with
respect to µF , and also σ CµF (A) = MµF (R). Combining obtains:
µ = µF , on MµF (R). (3)
We claim that:
MµF (R) = σ(P (R)), (4)
and this with (3) then proves (1).
To this end, recall the outer measure µ∗A defined on the power sigma algebra σ(P (R)).
By Remark I.5.17, for A ∈ σ(P (R)):
nX [ o
µ∗A (A) ≡ inf µA ((an , bn ]) | A ⊂ (an , bn ] ,
n n
0
where {(an , bn ]} ⊂ A can be taken to be disjoint. Now MµF (R) is the collection of
Carathéodory measurable sets with respect to µ∗A , MµF (R) ⊂ σ(P (R)) by definition. Also,
µ∗A is a measure on MµF (R), and we defined µF ≡ µ∗A .
Thus to prove (4) is to prove that every set A ∈ σ(P (R)) is Carathéodory measurable
with respect to µ∗A . That is, for any set E ⊂ R:
\ \
µ∗A (E) = µ∗A A E + µ∗A A e E .
If no such
S collection exists then we are done. Otherwise, by definition of µ there exists
{x
Pj } ⊂ n (an , bn ] − A, so choose xm with maximum µ ({xj }) from this collection. Since
j µ ({xk }) ≤ 1, max{µ ({xj })} is well defined and there can be at most finitely many xj
at this maximum, so choose any one.
Probability Spaces 13
This process can be continued for each such xm , of which there are at most countably
many. The infimum of these constructed collections converges to µ(A), since µ only has
positive measure on such xm . Thus (5) is proved, which proves (4). Combining with (3) we
obtain (1).
Three important and useful discrete probability measures are introduced here and
will facilitate examples below.
1. Discrete Rectangular Probability Measure
Perhaps the simplest probability measure that can be imagined is one which
assumes the same value on every one of a finite number of sample points. The do-
n
main of this distribution is arbitrary but conventionally taken as {xi } = {j/n}j=1
n
or sometimes {j/n}j=0 , so in either case {xi } ⊂ [0, 1]. For given n, the discrete
rectangular measure on [0, 1], sometimes called the discrete uniform mea-
n
sure, is defined on {j/n}j=1 by:
xi → (b − a)xi + a,
Exercise 1.10 (On nj )Show that nj equals the number of distinct j-element
subsets that can be chosen from a set of n distinct objects, j ≤ n, where by distinct
subset is implied that the ordering of elements is ignored.
The proof of the binomial theorem follows from the observation that the expansion
of (a + b)n is a sum of terms of the form {aj bn−j }, so it is only the coefficients
that need to be addressed. This coefficient identifies the number terms in this
product equal to aj bn−j . Now for given j, there are:
which is convergent for all x ∈ R. For more on Taylor series, see Chapter 9 of
Reitano (2010) or any book on calculus.
One important application of the Poisson distribution is that it provides a good
approximation to the binomial distribution when the binomial parameter p is
“small,” and the binomial parameter n is “large.” Specifically, with fixed λ = np,
then for p “small” and n “large”:
(np)j
n j
p (1 − p)n−j ' e−np . (1.13)
j j!
The requirement that p be “small” is typically understood as the condition that
p < 0.1, while n “large” is understood as n ≥ 100 or so.
The approximation in (1.13) was of critical value in pre-computer days, and comes
from the following result. Here it is assumed that as n increases, p decreases so
that the product np is fixed and equal to λ.
λj
n j
p (1 − p)n−j −→ e−λ , (1.14)
j j!
for all j as n → ∞.
Proof. First off:
j n −j
n j n(n − 1)...(n − j + 1) λ λ λ
p (1 − p)n−j = 1− 1−
j j! n n n
j n −j
n(n − 1)...(n − j + 1) λ λ λ
= j
1− 1− .
n j! n n
The second term, λj /j!, is fixed as a function of n and part of the final result, while the last
−j
term, 1 − nλ , converges to 1 as n → ∞ because the exponent −j is fixed. Similarly, the
Qj−1
first term equals the product of j-terms: k=0 (1 − nk ), which also converges to 1 as n → ∞.
n
The only subtlety here is to prove that for any real number λ, (1 − λ/n) −→ e−λ as
n → ∞. This is equivalent to proving that n ln(1 − λ/n) −→ −λ, and this follows from the
Taylor series expansion for ln(1 − x), which is convergent for −1 ≤ x < 1:
X∞
ln(1 − x) = − xj /j. (1.15)
j=1
Taylor series theory also allows a finite sum expression with remainder term:
ln(1 − x) = −x − ξ 2x /2,
Hence, ln(1 − λ/n) = −λ/n − ξ n /2, where 0 < ξ n < λ/n, so:
Exercise 1.12 Show that the limit n ln(1 − λ/n) −→ −λ also follows from the definition
of the derivative of f (x) = ln(1 − λx) at x = 0.
F (x ± ) → F (x),
3. By Chapter I.5, construct on R the Borel measure µF , and complete sigma algebra
MµF (R), producing the complete probability space denoted (R, MµF (R), µF ).
Equation (1.17) is the definition of F (x), while (1.16) follows from the fundamental theorem
of calculus.
Probability Spaces 17
In this model one can explicitly assign probabilities to right semi-closed intervals as in
Chapter I.5:
Z b
µF [(a, b]] = F (b) − F (a) = f (y)dy.
a
By continuity it follows that the same result is produced for closed, open and left semi-closed
intervals. From this it also follows that µ [{a}] = 0 for any point a.
Within the Riemann integration theory, while we know from Book I that the associated
µF is well defined on MµF (R) and thus also on B (R) ⊂ MµF (R), we cannot represent
µF [A] as above except for the simplest Borel sets, such as intervals. The reason for this is
simply that the associated Riemann integrals may not be defined.
Specifically, with χA (x) denoting the characteristic function of A, defined as χA (x) =
1 for x ∈ A and χA (x) = 0 otherwise, we will define:
Z Z
f (x)dx = f (x)χA (x)dx.
A
Thus within the Riemann theory, if f (x) is continuous, this integral will only exist if A is
a set for which χA (x) is continuous almost everywhere.
A simple substitution into the integral demonstrates that fN (x) and φ(x) have
the same Riemann integral over R. Also, for any N :
−N
exp −x2 /2 < |x|
, as |x| → ∞.
R∞
It follows that −∞ φ(y)dy < ∞ as an improper integral. However, there is no
√
elementary derivation that verifies that this function in fact integrates to 2π.
The simplest justification follows, and even this requires a deep result on multi-
variate integrals such as Fubini’s theorem, named for Guido Fubini (1879–
1943), or Tonelli’s theorem, named for Leonida Tonelli (1885–1946), as well
as results on substitution in integrals. These results will be derived in Book V,
and we assume these for now since in the current Riemann integral context these
assumptions are intuitively reasonable and are likely familiar to the reader.
Probability Spaces 19
x = r cos θ; y = r sin θ,
where r denotes the standard distance from (x, y) to the origin (0, 0), and θ the
angle in radian measure that the segment from (0, 0) to (x, y) makes with the
positive x-axis, so 0 ≤ θ < 2π.
This substitution will lead to dxdy = rdrdθ, and an integration over 0 ≤ r < ∞
and then 0 ≤ θ ≤ 2π obtains:
Z ∞ 2 Z 2π Z ∞
1
exp −r2 /2 rdrdθ = 1.
φ(y)dy =
−∞ 2π 0 0
Admittedly, this derivation requires a lot of faith that the various manipulations
are justifiable, but the reader will eventually see in Book V that this is so.
4. Lognormal Probability Measure
The probability density function of the lognormal probability measure is de-
fined on [0, ∞), depends on a location parameter µ ∈ R and a shape parameter
σ > 0, and unsurprisingly is intimately related to the normal probability measure
discussed above. However, to some the name “lognormal” appears to be opposite
of the relationship that exists. Stated one way, the variable x has a lognormal
probability function with parameters (µ, σ) if x = ez , where z has a normal prob-
ability function with the same parameters. So lognormal x can be understood as
an exponentiated normal. Stated another way, a variable x has a lognormal
probability function with parameters (µ, σ) if ln x has a normal probability func-
tion with the same parameters. The name comes from the second statement, in
that the log of a lognormal variable is normal.
The probability density function of the lognormal is defined on [0, ∞) as follows,
again using exp A ≡ eA to simplify notation:
1
fL (x) = √ exp − (ln x − µ)2 /2σ 2 , x ≥ 0, (1.25)
σx 2π
while fL (x) = 0 for x < 0. The substitution y = (ln x − µ) /σ produces:
Z ∞ Z ∞
1
exp −y 2 /2 dy
fL (x)dx = √
0 2π −∞
Z ∞
= φ(y)dy.
−∞
Thus, subject to the above “derivation,” the integral of the lognormal density
over [0, ∞) equals 1.
20 Foundations of Quantitative Finance
These events are said to be pairwise independent if (1.26) remains valid for any
subset J with k = 2.
Probability Spaces 21
for each of the 2N expressions where each Bn = An or Bn = S. See Proposition 1.19 for
another characterization.
1. General Binomial and independent coin flips: Recall the discussion above on
the general binomial measure µB defined on {0, 1, 2, ..., n} by (1.8). These probability
values were deemed to result from the summing of n “independent” standard binomials,
each of which takes value from {0, 1}.
Define the sample space S of n-tuples (x1 , x2 , ..., xn ) of results, each point of which
identifies the ordered outcome of n standard binomials. There are 2n distinct such
n-tuples. Define the 2n events:
where k = 0, 1, and note that Ai(k) is the collection of n-tuples with the ith component
specified to equal k. More visually, it is the collection of n-flips with the ith flip equal
to H if k = 1, or T if k = 0. The probability measure of these events is given by:
p, k = 1,
µ Ai(k) = (1)
p0 , k = 0,
where p0 ≡ 1 − p. This is because Ai(k) contains all n-tuples with xi = k and thus this
can be modeled as the full collection of (n − 1)-flips, which has probability 1, and then
an nth flip with probability p or p0 .
22 Foundations of Quantitative Finance
T
The events Ai(0) and Ai(1) cannot be independent for any i since A i(0) Ai(1) = ∅ and
thus (1.26) will not be satisfied since µ Ai(ki ) > 0. We now investigate the implica-
tions of more general independence of these sets since this independence characterizes
the independence of the standard binomials noted above. For example, that the ith and
jth coin flip are independent standard binomials for i 6= j, means that Ai(ki ) and Aj(kj )
are independent events for any choice of outcomes ki and kj .
More generally, given any collection {ki }ni=1 , the n events {Ai(ki ) }ni=1 are independent
if: \ n Yn
µ Ai(ki ) = µ Ai(ki ) .
i=1 i=1
Tn
Now i=1 Ai(ki ) = (k1 , k2 , ..., kn ), a single sample point, and so independence requires
by (1) that:
µ(k1 , k2 , ..., kn ) = pj (1 − p)n−j ,
where j denotes the number of indexes with ki = 1, and n − j the number with ki = 0.
Thus independence of {Ai(ki ) }ni=1 implies that the probability of (k1 , k2 , ..., kn ) depends
only on j, the number of indexes with ki = 1, and not the order.
There are nj of the Ai(k) -events that have exactly j indexes with ki = 1, and these are
mutually disjoint, meaning no sample point can be in any two of these events. Define
Bj as the union of all the sample points in these nj events. That is, Bj is the event
containing all sample points with j values equal to 1. It then follows by finite additivity
that:
n j
µ(Bj ) = p (1 − p)n−j .
j
This is equivalent to the probability measure in (1.8), and so µB (j) = µ(Bj ).
In other words, the probabilities assigned by (1.8) are consistent with the assumption
of n independent flips.
It is left as an exercise to show that given any collection {ki }m
i=1 for 2 ≤ m < n, the m
events {Ai(ki ) }m
i=1 are independent. Hint: Consider the argument demonstrating (1).
Next are a few simple yet useful results on independent events. The first addresses
“extreme” events A, meaning µ(A) ∈ {0, 1}.
Probability Spaces 23
Proposition 1.18 (Extreme events) If A ∈ E satisfies µ(A) ∈ {0, 1}, then A is inde-
pendent of every event B ∈ E including itself.
If A ∈ E is independent of itself, then µ(A) ∈ {0, 1}.
Proof. If µ(A) = 0, then A is independent of all events B since A ∩ B ⊂ A and hence
µ(A∩B) = 0 by subadditivity. If µ(A) = 1, then µ(A) = 0 and since B = (B ∩ A)∪ B ∩ A
e e ,
independence follows from finite additivity. Specifically, µ(B) = µ(A ∩ B), which is (1.26)
since µ(A) = 1.
2
If A is independent of itself, then (1.26) implies that µ(A) = (µ(A)) and hence µ(A) ∈
{0, 1}.
The next result provides an alternative characterization of independence.
Proposition 1.19 (Alternative characterization of independence) Let (S, E, µ) be
a probability space. Then {An }N
n=1 ⊂ E are independent events if and only if:
\ Y
N N
µ Bn = µ (Bn ) , (1.27)
n=1 n=1
Corollary 1.20 Let (S, E, µ) be a probability space, then {An }N n=1 ⊂ E are independent
by Definition 1.15 if and only if {Bn }N
n=1 are independent for each of the 2N choices where
Bn = An or Bn = A fn .
Proof. This follows from Proposition 1.19.
However, A12 ∩ A13 ∈ σ(A1 ), and this set is not independent of A23 ∈ σ(A2 ) since:
The problem here is that while A1 and A2 are independent classes, the sets A12 , A13 , A23
are only pairwise independent and not mutually independent:
The next result shows that if we begin with a collection of mutually independent sets and
partition this collection into classes, the above problem cannot occur and finite intersection
sets will again be independent. The most complicated part of this next result is the necessary
notation. So to simplify this, we represent the initial collection of mutually independent sets
as a matrix of sets which may have a finite or infinite number of rows, and each row a finite
or infinite number of elements. We will create classes by grouping rows of sets.
The goal of the next few results is to show that these classes generate independent sigma
algebras, and we do this in steps.
Proposition 1.22 (Independent classes) Given a probability space (S, E, µ), let A =
{Ajk } ⊂ E be a finite or infinite class of mutually independent sets with 1 ≤ j ≤ N and
1 ≤ k ≤ Mj where N and any Mj can be finite or infinite. Partition this class “by rows”
Mj 0
into subclasses {Aj }Nj=1 , with Aj = {Ajk }k=1 . Define Aj as the collection of sets from Aj
and all possible finite intersections of sets from Aj .
Then {A0j }Nj=1 are mutually independent classes.
Proof. First note that {Aj }N j=1 are independent classes since A is a collection of inde-
pendent sets and {Aj }N j=1 simply partitions this collection. In detail, given a finite subset
J ⊂ (1, ..., N ) with J ≡ {j1 , j2 , ...jn } and any {Aji ki }ni=1 with Aji ki ∈ Aji , then (1.26) is
satisfied because {Aji ki }ni=1 ⊂ A. Thus only the independence of the intersection sets from
{A0j }N
j=1 need be checked.
Probability Spaces 25
Choose {Bj } with Bj ∈ A0j for a finite index collection J = {j1 , ..., jn }, which by re-
T assume that J = {1, 2, ..., n}. Each Bj is
ordering rows and relabeling for simplicity we can
by assumption a finite intersection set so Bj = Kj Ajk with Ajk ∈ Aj , and where each Kj
is a finite index set which by re-ordering we can again assume that Kj = {1, 2, ..., kj }. Of
course, if kj = 1 for any j, then Bj = Aj1 .
Recalling that {Ajk }j,k are mutually independent:
\ n \
n \kj
µ Bj = µ Ajk
j=1 j=1 k=1
Yn Ykj
= µ (Ajk )
j=1 k=1
Yn \
kj
= µ Ajk
j=1 k=1
Yn
= µ (Bj ) .
j=1
Our final goal is to expand the conclusion of this corollary to assert that the smallest
sigma algebras generated by these classes, {σ(Aj )}N
j=1 , are mutually independent classes.
We then extend this result.
Proposition 1.25 (Independent sigma algebras 1) Given a probability space (S, E, µ),
let A = {Ajk } ⊂ E be a finite or infinite class of mutually independent sets with 1 ≤ j ≤ N
Mj
and 1 ≤ k ≤ Mj , partitioned into subclasses {Aj }N j=1 with Aj = {Ajk }k=1 .
26 Foundations of Quantitative Finance
If σ(Aj ) denotes the smallest sigma algebra generated by Aj , then {σ(Aj )}N j=1 are mu-
tually independent.
Proof. Consider a finite collection of semi-algebras, say {A0 (Aj )}nj=1 by relabeling indexes,
and which are mutually independent classes by corollary 1.23. Thus given Aj ∈ A0 (Aj ):
\n Yn
µ Aj = µ (Aj ) .
j=1 j=1
Then A0 (A1 ) ⊂ C(A1 ), and we now show that C(A1 ) is a sigma algebra.
First, C(A1 ) is closed under complementation by Proposition 1.19, so A ∈ C(A1 ) if and
only if Ae ∈ C(A1 ). Next, consider a countable collection, {A1k }∞ ⊂ C(A1 ) which are
T S∞ k=1
disjoint, A1k A S1i = ∅ for k 6= i, and we show that k=1 A1k ∈ C(A1 ). To this end, note
∞
that with A1 ≡ k=1 A1k , it follows from De Morgan’s laws that:
\n [∞ h \ \ n i
Aj = A1k Aj ,
j=1 k=1 j=2
n T Tn o∞
and since {A1k }∞
k=1 are disjoint, so too are A 1k j=2 A j .
k=1
Thus by countable additivity:
\ n X∞ h \ \ n i
µ Aj = µ A1k Aj
j=1 k=1 j=2
X∞ h Yn i
= µ (A1k ) µ (Aj )
k=1 j=2
Yn
= µ (Aj ) .
j=1
S∞
Now consider A1 ≡ k=1 A1k where {A1k }∞ k=1 ⊂ C(A S1 )∞are not necessarily
S∞ disjoint.
In this case there exists disjoint {A01k }∞
k=1 ⊂ E for which A
k=1 1k = 0
k=1 1k For this
A .
construction, define A011 = A11 , and then for k > 1:
\ \k−1
A01k = A1k A
e1j .
j=1
As each A1k ∈ C(A1 ), so too is each A e1k ∈ C(A1 ) by (1.27), and thus as a finite intersection,
0
each A1k ∈ C(A1 ) by Proposition 1.22. Applying the conclusion with disjoint sets and noting
that: X∞
µ (A01k ) = µ (A1 ) ,
k=1
S∞
by countable additivity, we conclude that k=1 A1k ∈ C(A1 ).
Hence C(A1 ) is a sigma algebra that contains A1 , and thus by definition contains the
smallest such sigma algebra σ(A1 ). Now the definition of C(A1 ) depended on the initial
selection of {Aj }nj=2 , but for any selection the conclusion is the same, that σ(A1 ) ⊂ C(A1 ),
and {σ(A1 ), A0 (A2 ), ..., A0 (An )} are mutually independent classes.
We can repeat this construction iteratively, next investigating C(A2 ) by fixing A1 ∈
σ(A1 ) and Aj ∈ A0 (Aj ) for 3 ≤ j ≤ n, and proving σ(A2 ) to be independent of the other
classes. Repeating, we finally conclude that {σ(Aj )}nj=1 are mutually independent classes.
Since this is then true for any finite collection, {σ(Aj )}N j=1 are mutually independent classes
by Definition 1.15.
Another random document with
no related content on Scribd:
The New Baby
(Written for Alfred Greene, Jr., Evansville, Indiana.)
When Alfred saw the baby wee the stork to him had
brought,
"Dear mother we must send this kid back to her home to-
day,
Then you and I'll be glad again and go to hear the band.
Woes Caused By Whooping Bugs
(Written for Cousin Harvey Stoner, Jr.)
And even when you're tired you can't sleep for the croup,
But you must stay at home ALONE from three to six long
weeks,
Said young Heiman to Simple Simon, "I will not selfish be,
There were bisque dolls and wax dolls and dolls with real
hair,
Red dolls and black dolls and dolls that were fair,
But though I love dear mother far more than all the rest,
And mother says perhaps this solves the very reason why
A la mode to be attired.
But when she rose on Easter morn
But one good giant's life was spared by this bold warrior
lad.
ARITHMOS was this giant great, and all bright girls and
boys
Should love the famous Giant-King far more than all their
toys.
He's very old, and very great, and also wondrous wise,
For he can count all things on earth and even tell their
size.
He knows how many birds there are; how high each bird
can fly,
'Tis he alone can tell you when a great eclipse will come
And darken the moon's lady or the old man in the sun.
He'll help us with our lessons hard when for his aid we
call,
And for his work and trouble never thinks to ask a penny.
We all should learn his wondrous truths and love him more
and more.
Good Weather Assured
(Written for the Evansville Courier, February 3, 1909.)
And eat farmers' crops 'neath the light o' the sun.
One day last week King Teddy arose with old King Sun,
But when Queen Lion heard his tale, she simply scratched
one ear,
Then shrugged her shoulders a la hump and to her
husband said,
Then brave and proud Queen Lion she carried out each
babe,
For of his teeth and of his gun she hadn't any fear.