Professional Documents
Culture Documents
Random Vectors: An Overview: Master INVESTMAT 2018-2019 Unit 2
Random Vectors: An Overview: Master INVESTMAT 2018-2019 Unit 2
Random Vectors: An Overview: Master INVESTMAT 2018-2019 Unit 2
In this session we will remember the main definitions, concepts, properties and results
related to multivariate random variables, or equivalently, random vectors. The
exposition follows on purpose, in broad outline, the same structure as univariate r.v.’s
since most of the ideas are extensions of their unidimensional counterpart. In this way,
we pursue clarity rather than appealing.
To be equipped with a good background of random vectors is crucial in dealing with
systems of random differential equations (R.D.E.’s) since their inputs, such as, initial
and/or boundary conditions, source terms and coefficients can be random vectors
and/or matrices rather than deterministic vectors and/or matrices, respectively.
Random Vectors
lim FX1 ,X2 (x1 , x2 ) = lim FX1 ,X2 (x1 , x2 ) = lim FX1 ,X2 (x1 , x2 ) = 0.
(x1 ,x2 )→(−∞,−∞) x1 →−∞ x2 →−∞
N M
FX1 (x1 ) = ∑ FX1 ,X2 (x1,i , x2,j ), FX2 (x2 ) = ∑ FX1 ,X2 (x1,i , x2,j ).
j=1 i=1
Continue r.v.’s:
Z ∞ Z ∞
FX1 (x1 ) = FX1 ,X2 (x1 , x2 )dx2 , FX2 (x2 ) = FX1 ,X2 (x1 , x2 )dx1 .
−∞ −∞
lim FX1 ,X2 (x1 , x2 ) = FX1 (x1 ), lim FX1 ,X2 (x1 , x2 ) = FX2 (x2 ).
x2 →∞ x1 →∞
In the n–dimensional case, and assuming that X = (X1 , . . . , Xn ) is a continuous random vector the i-th marginal d.f.
is given by: Z ∞ Z ∞
FX (xi ) = ··· FX1 ,...,Xn (x1 , . . . , xn )dx1 · · · dxi−1 dxi+1 · · · dxn , 1 ≤ i ≤ n.
i −∞ −∞
It satisfies that:
FX1 ,X2 (x1 , x2 ) = ∑ ∑ pX1 ,X2 (x1,i , x2,j ).
x1,i ≤x1 x2,j ≤x2
If X1 and X2 are continuous r.v.’s, the joint probability density function (p.d.f.) is a
function fX1 ,X2 (x1 , x2 ) such that:
2
Z Z x1 x2 fX1 ,X2 (x1 , x2 ) ≥ 0, ∀(x1 , x2 ) ∈ R ,
FX1 ,X2 (x1 , x2 ) = fX1 ,X2 (x1 , x2 ) dx1 dx2 , Z ∞Z ∞
−∞ −∞
fX1 ,X2 (x1 , x2 ) dx1 dx2 = 1.
−∞ −∞
In addition, the following relationships between marginal p.d.f.’s and joint p.d.f. hold:
Z ∞ Z ∞
fX1 ,X2 (x1 , x2 ) dx2 = fX1 (x1 ), fX1 ,X2 (x1 , x2 ) dx1 = fX2 (x2 ).
−∞ −∞
In addition, the following relationship between the joint p.d.f. and joint d.f. are fulfilled
∂ 2 FX1 ,X2 (x1 , x2 )
fX1 ,X2 (x1 , x2 ) = .
∂ x1 ∂ x2
We do not rewrite the above relationships in the multi–dimensional case.
Random Differential Equations and Applications Random Vectors: An overview 6
Example 1: Computing some significant distributions from the joint p.d.f.
Let (X , Y ) be a two-dimensional r.v. with joint p.d.f.
y 2
f (x, y ) = 6 x − , 0 ≤ x ≤ 1, 0 ≤ y ≤ 1.
2
indep. P [E1 ∩ E2 ]
P [E1 ] = P [E1 |E2 ] = ⇒ P [E1 ∩ E2 ] = P [E1 ] P [E2 ] .
P [E2 ]
FX1 ,X2 (x1 , x2 ) = P [A1 ∩ A2 ] = P [A1 ] P [A2 ] = FX1 (x1 )FX2 (x2 ).
X1 , X2 are independent r.v.’s ⇔ FX1 ,X2 (x1 , x2 ) = FX1 (x1 )FX2 (x2 ).
The definition can also be given by discrete r.v.’s with m.p.f. pX1 ,X2 (x1 , x2 ).
As a consequence one gets the following characterization (or definition) of
independence of two r.v.’s:
X1 , X2 are independent r.v.’s ⇔ pX1 |X2 (x1 |x2 ) = pX1 (x1 ) and pX2 |X1 (x2 |x1 ) = pX2 (x2 ).
fX1 ,X2 ,X3 (x1 , x2 , x3 ) = fX1 |X2 ,X3 (x1 |x2 , x3 )fX2 |X3 (x2 |x3 )fX3 (x3 ).
X1 , X2 , X3 are mutually independent r.v.’s ⇔ fX1 ,X2 ,X3 (x1 , x2 , x3 ) = fX1 (x1 )fX2 (x2 )fX3 (x3 ).
fX1 ,...,Xn (x1 , . . . , xn ) = fX1 |X2 ,...,Xn (x1 |x2 , . . . , xn )fX2 |X3 ,...,Xn (x2 |x3 , . . . , xn ) · · · fXn−1 |Xn (xn−1 |xn )fXn (xn ),
3(y − 2x)2
fX |Y (x|y ) = , 0 ≤ x ≤ 1, 0 ≤ y ≤ 1.
4 + 3y (y − 2)
3(y − 2x)2
fY |X (y |x) = , 0 ≤ x ≤ 1, 0 ≤ y ≤ 1.
1 + 6x(2x − 1)
Check that both are p.d.f.’s and show a plot of them.
P[A ∩ B] = P[A]P[B],
P[A ∩ C ] = P[A]P[C ], (1)
P[B ∩ C ] = P[B]P[C ],
and
P[A ∩ B ∩ C ] = P[A]P[B]P[C ]. (2)
It is possible that relationship (2) holds, but some of the relationships fails (1).
Conversely, next example shows that (1) holds true but (2) fails.
1
Ω = {s1 , s2 , s3 , s4 }, P[si ] = , i = 1, . . . , 4.
4
Then
A ∩ B = A ∩ C = B ∩ C = A ∩ B ∩ C = {s1 },
1
= = P[C ] = ,
P[A] P[B]
2
1
P[A ∩ B]
= P[A ∩ C ] = P[B ∩ C ] = P[A ∩ B ∩ C ] = .
4
Then
3 3 1
fX1 ,X2 (0.5, 0.5) = 2 6= = × = fX1 (0.5)fX2 (0.5) ⇒ fX1 ,X2 (x1 , x2 ) 6= fX1 (x1 )fX2 (x2 ),
4 2 2
D = {(x1 , x2 ) ∈ R2 : 0 ≤ x1 ≤ x2 ≤ 1} 6= D1 × D2 ,
i.e., the domain where the joint p.d.f. fX1 ,X2 (x1 , x2 ) is defined is not a product space
(plot the domain!).
Remark: A necessary condition of independence is the domain D be the product of
several spaces. However, this condition is not sufficient (see Exercise 2 (3)).
Random Differential Equations and Applications Random Vectors: An overview 16
From the concept of conditional distribution one can introduce the definition of
truncated distribution which is very useful in practice where often the support of the
r.v.’s is bounded.
Truncated distributions
Let X be a real r.v. defined on a probability space (Ω, FΩ , P) and T ∈ BR such that:
0 < P[{ω ∈ Ω : X (ω) ∈ T }] < 1. Then
Discrete r.v.: If X is a discrete r.v. with m.p.f. pX (x) = P[X = x], the truncated
m.p.f. over the set T is given by:
pX (x)
P[X = x, X ∈ T ]
if x ∈ T ,
P[X = x|X ∈ T ] = = ∑ pX (t)
P[X ∈ T ]
t∈T
0 otherwise.
Continuous r.v.: If X is a continuous r.v. with p.d.f. fX (x), the truncated p.d.f.
over the set T is given by:
f (x)
Z X
if x ∈ T ,
fX |T (x|X ∈ T ) = fX |T (x) = fX (y ) dy
T
0 otherwise.
10 x 10−x
x p (1 − p)
pX (x) = 6 , x = 0, 1, . . . , 6.
10
∑ x px (1 − p)10−x
x=0
e −λ x
fX (x) = Z 20
, 0 ≤ x ≤ 20.
e −λ y dy
0
In particular:
Moment w.r.t. the origin: Taking g (X1 , X2 ) = (X1 )m (X2 )n , m, n ≥ 1, one gets:
Z ∞Z ∞
αm,n = E[(X1 )m (X2 )n ] = (x1 )m (x2 )n fX1 ,X2 (x1 , x2 ) dx1 dx2 .
−∞ −∞
Moment w.r.t. the mean: Taking g (X1 , X2 ) = (X1 − µX1 )m (X2 − µX2 )n , m, n ≥ 1,
one gets:
Z ∞Z ∞
µm,n = E[(X1 − µX1 )m (X2 − µX2 )n ] = (x1 − µX1 )m (x2 − µX2 )n fX1 ,X2 (x1 , x2 ) dx1 dx2 .
−∞ −∞
Remarks:
1 The above definitions can also be stated for discrete r.v.’s X1 and X2 substituting integrals by sums and
the joint p.d.f. by the joint p.m.f.
2 In the same way as we did for univariate r.v.’s, one can defined the absolute moments w.r.t the origin and
the mean.
3 These definitions can also be given for n–dimensional random vectors.
Correlation (α1,1 ):
4
α1,1 = E[X1 X2 ].
4
µ1,1 = E[(X1 − µX1 )(X2 − µX2 )] = α1,1 − α1,0 α0,1 = E[X1 X2 ] − E[X1 ]E[X2 ].
4 µ1,1
ρX1 ,X2 = √ ∈ [−1, 1] (by Schwarz inequality).
µ2,0 µ0,2
⇒
X1 , X2 are independent r.v.’s E[f (X1 )g (X2 )] = E[f (X1 )]E[g (X2 )].
:
σX 2
a = ρX1 ,X2 , b = µX2 − aµX1 , e = σX22 (1 − (ρX1 ,X2 )2 ).
σX1
As a consequence:
σX2
a= > 0,
σX1
ρX1 ,X2 = 1 ⇒ X2 = aX1 + b,
σ
b = µX2 − σX2 µX1 .
X1
e = 0 ⇔ |ρX1 ,X2 | = 1 ⇒
a = − σX2 < 0,
X1 σ
ρX1 ,X2 = −1 ⇒ X2 = aX1 + b, b = µ + σX2 µ .
X2 σ X1 X1
This is the case where linear approximation is the best in the m.s. sense. The opposite
case is when the error is maximum, i.e.,
= P[X 2 ≤ u, Y 2 ≤ v ]
√ √
= P[X ≤ u, Y ≤ v ]
F √ √
= P[X ≤ u]P[Y ≤ v ]
= P[X 2 ≤ u]P[Y 2 ≤ v ]
= P[U ≤ u]P[V ≤ v ]
= FU (u)FV (v ).
The first part of the previous exercise shows that two r.v.’s which are functionally
dependent can be statistically independent.
If X1 , X2 ,. . . , Xn are mutually independent r.v.’s (or milder, uncorrelated), then
C Xi , Xj = 0 if i 6= j and previous expression yields
" #
n n
V ∑ a i Xi = ∑ ai2 V [Xi ] .
i=1 i=1
σX2
C[X1 , X1 ] C[X1 , X2 ] ρX1 ,X2 σX1 σX2
1
Σ X,X = = .
C[X2 , X1 ] C[X2 , X2 ] ρX1 ,X2 σX1 σX2 σX2
2
4 µ1,1 C[X1 , X1 ]
ρX1 ,X2 = √ = .
µ2,0 µ0,2 σX1 σX2
x1 − µX1 x2 − µX2
− 2ρ
σX 1 σX 2
2 #)
x2 − µX2
+ .
σX 2
µ X ; ΣX ), A ∈ Rm×n and b ∈ Rm .
H : Let X = (X1 , . . . , Xn ) ∼ N(µ
µ X + b; AΣX AT ).
T : Then AX + b ∼ N(Aµ
We showed that
⇒
X1 , X2 are independent r.v.’s X , X uncorrelated r.v.’s,
: 1 2
however
The definition in the case that both r.v.’s are discrete can be made in analogous way.
This complex function always exists.
ϕX1 ,X2 (u1 , u2 ) is just the two–dimensional Fourier transform the joint p.d.f.
fX1 ,X2 (x1 , x2 ). By the Fourier inversion theorem, one can recover the joint p.d.f.:
1
Z ∞Z ∞
fX1 ,X2 (x1 , x2 ) = e −i(u1 x1 +u2 x2 ) ϕX1 ,X2 (u1 , u2 ) du1 du2 .
4π 2 −∞ −∞
H : Let X = (X1 , X2 ) be a two-dimensional r.v. with joint p.d.f. fX1 ,X2 (x1 , x2 ). Let
r : R2 → R2 a one-to-one deterministic map and s : R2 → R2 its inverse:
y1 = r1 (x1 , x2 ), x1 = s1 (y1 , y2 ),
y2 = r2 (x1 , x2 ), x2 = s2 (y1 , y2 ).
Let us assume that both maps are differentiable being their four partial derivatives
continuous. Let us also assume that the Jacobian J2 of the inverse map satisfies:
∂ x1 ∂ x2
!
∂ y1 ∂ y1
J2 = det ∂ x1 ∂ x2 6= 0.
∂ y2 ∂ y2
T : Then, the joint p.d.f. fY1 ,Y2 (y1 , y2 ) of the two-dimensional r.v.
Y = (Y1 , Y2 ) = (r1 (X1 , X2 ), r2 (X1 , X2 )) is given by
where s(y) is the inverse transformation of r(x): x = r−1 (y) = s(y) and Jn is the
Jacobian of the transformation, i.e.,
∂ x1 ∂ xn
∂ y1 ··· ∂ y1
∂x .. .. ..
Jn = det = det . . .
,
∂y
∂ x1 ∂ xn
∂ yn ··· ∂ yn
Y1 = r1 (X1 , X2 ) = X1 + X2 , Y2 = r2 (X1 , X2 ) = X1 ,
H : Let (X1 , X2 ) be a continuous random vector with joint p.d.f. fX1 ,X2 (x1 , x2 ) and
respective domains: DX1 = {x1 : x1,1 ≤ x1 ≤ x1,2 } and DX2 = {x2 : x2,1 ≤ x2 ≤ x2,2 }.
T : Then the p.d.f. fY1 (y1 ) of their sum Y1 = X1 + X2 is given by:
Z x1,2
fY1 (y1 ) = fX1 ,X2 (x1 , y1 − x1 )dx1 , y1,1 = x1,1 + x2,1 ≤ y1 ≤ x1,2 + x2,2 = y1,2 ,
x1,1
or, equivalently by
Z x2,2
fY1 (y1 ) = fX1 ,X2 (y1 − x2 , x2 )dx2 , y1,1 = x1,1 + x2,1 ≤ y1 ≤ x1,2 + x2,2 = y1,2 .
x2,1
If X1 and X2 are independent r.v.’s, the p.d.f. of the sum of two independent r.v.’s is
just the convolution of their respective p.d.f.’s:
Z x1,2 Z x2,2
fY1 (y1 ) = fX1 (x1 )fX2 (y1 − x1 )dx1 , or fY1 (y1 ) = fX1 (y1 − x2 )fX2 (x2 )dx2 .
x1,1 x2,1
copula
A copula is a d.f.
C : [0, 1]n −→ [0, 1]
of n dependent uniform r.v.’s U1 , . . . , Un ∼ U([0, 1]):
where Tν,Σ is the d.f. of the multivariate t0,Σ distribution, with ν degrees of freedom,
mean vector 0 and correlation matrix Σ (i,e., Σ is a covariance matrix with ones in the
main diagonal), and Tν−1 is the inverse of the d.f. of the univariate tν distribution.
This includes the special case (ν = ∞) of the Gaussian copula model. In this setting
the dependency structure in the random vector X is determined by the correlation
matrix Σ.
n
C (u1 , . . . , un ) = ∏ ui .
i=1
Let Xi ∈ LRV = L2 , 1 ≤ i ≤ n and consider the set Ln2 whose elements are n-dimensional
random vectors X = (X1 , . . . , Xn ). This set has an structure of linear space and with
the norm:
kXkLn2 = max kXi kRV ,
1≤i≤n
is a Banach space. Therefore the concept of convergence in this space is the one
inferred by the above norm.