Professional Documents
Culture Documents
Week3 Lecture1 PDF
Week3 Lecture1 PDF
Semester 191
Week-3, Lecture-1
1
Functions of Random Variables (Ch. 5, 6)
2
Functions of Random Variables (Ch. 5, 6)
aX b
sin X X2
1
X Y g( X ) |X |
X
log X
eX | X | U ( x)
3
Y
a 0
y
Functions of Random Variables (Ch. 5, 6)
Y y
0 X
x
Example 1: Y aX b X
y b
a
Solution: Suppose a 0.
y b yb
FY ( y ) P Y ( ) y P aX ( ) b y P X ( ) FX .
a a
1 yb
fY ( y ) f X .
and a a Y
y b y
FY ( y ) P Y ( ) y P aX ( ) b y P X ( )
a Y y X
y b 0 x
1 FX ,
a 1 y b y b
f (y ) f X
1 yb
Y X
|a | a a
and hence fY ( y ) f X . 4
a a
Y X2
y
Example 2: Y X 2.
FY ( y ) PY ( ) y P X 2 ( ) y .
5
Y X2
y
Hence
FY ( y ) P x1 X ( ) x2 FX ( x2 ) FX ( x1 )
FX ( y ) FX ( y ), y 0.
6
Functions of Random Variables (Ch. 5, 6)
1
In particular if X N (0,1), so that f X ( x) e x / 2 ,
2
2
and substituting this , we obtain the p.d.f of Y X 2 to be
1
fY ( y ) e y / 2U ( y ).
2y
7
Functions of Random Variables (Ch. 5, 6)
Note: As a general approach, given Y g ( X ), first sketch the graph
y g ( x ), and determine the range space of y. Suppose a y b is
the range space of y g ( x ).Then clearly for y a, FY ( y ) 0, and for
y b, FY ( y ) 1, so that FY ( y ) can be nonzero only in a y b. Next,
determine whether there are discontinuities in the range space of y. If
so evaluate PY ( ) yi at these discontinuities. In the continuous
region of y, use the basic approach FY ( y) Pg ( X ( )) y
and determine appropriate events in terms of the r.v. X for every y.
Finally, we must have FY ( y ) for y , and obtain
dFY ( y )
fY ( y ) in a y b.
dy 8
Functions of Random Variables (Ch. 5, 6)
x
x1 x1 x1 x3 x3 x3
x2 x2 x2
y y
Py Y ( ) y y fY (u )du fY ( y ) y.
y
and as y 0, fY ( y ) 1
f X ( xi )
1
f X ( xi ).
i dy / dx x i g ( xi )
i
fY ( y ) 2 y X X
0, otherwise , 10
Mean, Variance, Moments and Characteristic Functions
For a r.v. X, its p.d.f f X (x) represents complete information about it,
and for any Borel set B on the x-axis
P X ( ) B f X ( x )dx.
B
Note that f X (x) represents very detailed information, and quite often it
is desirable to characterize the r.v. in terms of its average behavior. In
this context, we will introduce two parameters - mean and variance -
that are universally used to represent the overall properties of the r.v.
and its p.d.f.
11
Mean, Variance, Moments and Characteristic Functions
Mean or the Expected Value of a r.v. X is defined as
X X E( X ) x f X ( x )dx.
xi pi xi P ( X xi ) .
i i
Mean represents the average (mean) value of the r.v. in a very large
number of trials. For example if X U ( a, b), then
2 b
b x 1 x b2 a 2 ab
E( X )
a ba
dx
ba 2
2(b a )
2
a
k!
e
k
k 1 k!
k
i
e
(k 1)! e i!
k 1
i 0
e e .
If X is binomial, then
n n
n k n k n
n!
E ( X ) kP( X k ) k p q k p k q n k
k 0 k 0 k k 1 (n k )!k!
n
n! n 1
(n 1)!
p q np
k n k
pi q n i 1 np( p q)n 1 np.
k 1 ( n k )!( k 1)! i 0 ( n i 1)!i!
13
Mean, Variance, Moments and Characteristic Functions
1 1
ye
y 2 / 2 2 y 2 / 2 2
dy e dy .
2 2
2
2
0
1
Given X f X ( x), suppose Y g ( X ) defines a new r.v. with pdf
fY ( y ). Then the new r.v. Y has a mean Y given by
Y E (Y ) y fY ( y )dy.
k 0 k 0 k! k 1 k!
k
i 1
e
k (k 1)! e (i 1)
k 1
i 0 i!
i
i
i
e i
e i e
i 0 i! i 0 i! i 1 i!
i
m 1
e
e e e
i 1 (i 1)! m 0 m!
e e e 2 .
Mean alone will not be able to truly represent the p.d.f of any r.v. To
illustrate this, consider the following scenario: Consider two
Gaussian r.v.s X1 N (0,1) and X 2 N (0,10).Both of them have the
same mean 0. However, as Fig. shows, their p.d.f. s are quite
different. One is more concentrated around the mean, whereas the
other one ( X 2 ) has a wider spread. Clearly, we need at least an
additional parameter to measure this spread around the mean!
f X 1 ( x1 ) f X 2 ( x2 )
x1 x2
(a) 1
2
(b) 2 10
16
Mean, Variance, Moments and Characteristic Functions
For a r.v. X with mean , X represents the deviation of the r.v.
from its mean. Since this deviation can be either positive or negative,
consider the quantity X 2 , and its average value E [ X 2 ]
represents the average mean square deviation of X around its mean.
Define 2 E[ X ] 0. With g ( X ) ( X )2 and using
2
X
E (Y ) E g (X ) y f Y ( y )dy g (x )f X (x )dx ,
( x )2 f X ( x )dx 0.
2
X
x 2
2 x 2 f X ( x )dx
x f X ( x )dx 2 x f X ( x )dx 2
2
E X E X E ( X ) 2
___
X X .
2
2 2 2 2
X X 2 2 .
___
2
2 2
X
Thus for a Poisson r.v, mean and variance are both equal to its
parameter .
18
Mean, Variance, Moments and Characteristic Functions
Variance of the normal r.v. N ( , 2 )
x
1
Var( X ) E [( X ) ] ( x ) 2 / 2 2
2 2
e dx.
2 2
3
1
x ( x )2 / 2 2
dx 2 .
2
or e
2 2
*In some cases, mean and variance may not exist. For example, Cauchy r.v. 19
Mean, Variance, Moments and Characteristic Functions
___
20
Mean, Variance, Moments and Characteristic Functions
Characteristic Function
The characteristic function of a r.v. X is defined as
X ( ) E e e jx f X ( x )dx.
jX
Thus (0) 1, and ( ) 1 for all .
X X
Characteristic Function
Similarly, if X is a binomial r.v., its characteristic function is
given by
n
n k n k n n
X ( ) e jk
p q ( pe j ) k q n k ( pe j q) n .
k 0 k k 0 k
22
Mean, Variance, Moments and Characteristic Functions
Characteristic Function
Characteristic functions are useful in computing the moments
of a r.v. ( ) E e jX E ( jX )k j k E ( X k ) k
X
k 0 k! k 0 k!
E( X 2 ) 2 k
k E( X )
1 jE ( X ) j2
j k .
2! k!
Taking the first derivative with respect to ω, and letting it to
be equal to zero, we get
X ( ) 1 X ( )
jE ( X ) or E ( X ) .
0 j 0
Similarly, second derivative gives
1 2 X ( )
E( X ) 2
2
,
j 2
0 23
Mean, Variance, Moments and Characteristic Functions
Characteristic Function
and repeating this procedure k times, we obtain the kth moment of X
to be 1 k ( )
E( X k ) X
, k 1.
j k
k
0
We can use this technique to compute the mean, variance and other
higher order moments of any r.v. X. For example, if X P ( ), then
X ( ) e , X ( ) e ee je j , and
(e j 1) j
2 X ( )
2
e
e e
( je j 2
)
j
e e
j 2 j
e .
j
so that E ( X ) , and E ( X 2 ) 2 ,
24
X
X
Chebychev Inequality 2
25
P| X | ?
X
X
Chebychev Inequality 2
2 f X ( x )dx 2 f X ( x )dx 2 P | X | .
|x | |x |
P | X | k 2 .
1
k we obtain
k 26
P| X | ?
Chebychev Inequality
Thus with k 3, we get the probability of X being outside the 3
interval around its mean to be 0.111 for any r.v. Obviously this cannot
be a tight bound as it includes all r.v.s. For example, in the case of a
Gaussian r.v., from table of error function, ( 0, 1)
P | X | 3 0.0027.
which is much tighter than that given by Chebychev inequality
Chebychev inequality always underestimates the exact probability.
27