Professional Documents
Culture Documents
PQT-UNIT-3
PQT-UNIT-3
PQT-UNIT-3
DEPARTMENT OF MATHEMATICS
CLASS NOTES
MA 8402 – PROBABILITY AND QUEUEING THEORY
UNIT – III RANDOM PROCESSES
Markov process and Markov chains
Random process
A random process is an infinite indexed collection of random variables { X(t) : t T }, defined over a
common probability space. The index parameter t is typically time, but can also be a spatial dimension.
Random processes are used to model random experiments that evolve in time such as Daily price of a
stock.
If the time(parameter) ‘t’ is fixed, random process is called a random variable.
Continuous-time random process and Discrete state random process.
Consider a random process {X(t), t T}, where T is the index set or parameter set. The values assumed by
X(t) are called the states, and the set of all possible values of the states forms the state space E of the
random process.
(a) If the state space E and index set T are both continuous, then the random process is called
continuous-time random process.
(b) If the state space E is discrete and the index set T is continuous, then the random process is called
discrete state random process
Stationary process
A random process is said to be stationary if its mean ,variance , moments are constants other process are
called non stationary
Evolutionary process: A random process is not stationary in any sense is called as evolutionary process
First order stationary process:
A random process is said to be stationary to order one if is first order density function does not change with a shift in
time origin.i.e., f X x1 : t1 f X x1 , t1 for any time t1 and any real number .i.e., E X t X Constant.
second order stationary process:
A random process is said to be stationary to order two if its second-order density functions does not change with a
shift in time origin.i.e., f X x1 , x2 : t1 , t2 =f X x1 , x2 : t1 , t2 for all t1 , t2 and .
nth order stationary process, when will it become a SSS process?
A random process X t is said to be stationary to order n or nth order stationary if its nth order density
function is invariant to a shift of time origin.
f X x1 , x2 ,..., xn , t1 , t2 ,..., tn f X x1, x2 ,..., xn , t1 , t2 ,..., tn for all t1 , t2 ,..., tn & h . nth order
stationary process becomes a SSS process when n
Ergodic Random process.
A random process {X(t)} is called ergodic if all its ensemble averages are equal to appropriate
time averages
Autocorrelation
RXX t1 , t2 E X t1 X t2
Auto covariance
CXX t1 , t2 RXX t1 , t2 E X t1 E X t2
A random process {X(t)} is called wide-sense stationary if the following conditions hold:
(i) EX t a constant
(ii) RXX t1 , t 2 EX t1 X t 2 RXX t1 t 2 = function of time difference
Auto correlation:
RXX t1 , t 2 EX t1 X t 2 RXX t1 t 2
Evolutionary process:
A random process is not stationary in any sense is called as evolutionary process
which shows first order stationary random process has a constant mean.
2.Define a kth order stationary process. When will it become a strict sense stationary proess
(APR 2017)
The process is said to be kth order stationary if all its finite dimensional distributions are invariant under
translation of time for all t1,t2,….. tn and forn=1,2,3……k only and not for all n>k, then the process is called
the kth order stationary process. It becomes a strict sense stationary proess when k→∞.
3. Show that the random process X (t ) A cosw0 t is wide-sense stationary, if A and w0 are
constants and is uniformly distributed random variable in (0,2) (APR 2017)
Solution:
1
Since is uniformly distributed in 0, 2 p.d.f. is f 0 0 2
2
2
1 A A
E X t A cos 0t d sin 0t 0 sin 0t 2 sin 0t
0
2 2 2
A
sin 0t sin 0t 0
2
sin t t 2 2
A2 2
cos 0 t1 t2 0
0 1 2
4
2 0
sin 0 t1 t2 4 sin 0 t1 t2
A2
cos 0 t1 t2 2
4
2 2
A sin 0 t1 t2 sin 0 t1 t2
2
2 cos 0 t1 t2
4 2 2
2 2
A A
.2 cos 0 t1 t2 cos 0 t1 t2 =a function of t1 t2
4 2
The process X t is W.S.S.
4.Show that the process X (t ) A cos t B sin t is wide sense stationary, if E(AB) =0; E(A)=E(B) =
0, & E(A2) = E(B2), where A &B are random variables. (MAY 2013)
Solution:
Given X (t ) A cos t B sin t , E(A) = E(B) = E(AB) = 0, E(A2) = E(B2) = k(say)
To prove X(t) is wide sense stationary, we have to show
(i) E{X(t)} = constant
(ii) R XX (t,t+τ) = function of timedifference = function of τ
Now, E X t E A cos t B sin t cos tE ( A) sin tE ( B) 0 E X t constant
R XX (t , t ) E X (t ) X (t )
= E A cos t B sin t A cos t B sin t
A2 cos t cos t AB cos t sin t AB sin t cos t
= E
B 2
sin t sin t
= cos t cos t E A2 cos t sin t E AB sin t cos t E AB
sin t sin t E B 2
= cos t cos t k sin t sin t k
E ( AB) 0 & E ( A2 ) E ( B 2 ) k ( say)
= k cos t cos t sin t sin t
Solution:
Given X (t ) A cos t B sin t , E(A) = E(B) = E(AB) = 0, E(A2) = E(B2) = k(say)
To prove X(t) is wide sense stationary, we have to show
(i) E{X(t)} = constant
(ii) R XX (t,t+τ) = function of timedifference = function of τ
Now, E X t E A cos t B sin t cos tE ( A) sin tE ( B) 0 E X t constant
R XX (t , t ) E X (t ) X (t )
= E A cos t B sin t A cos t B sin t
A2 cos t cos t AB cos t sin t AB sin t cos t
= E
B 2 sin t sin t
= cos t cos t E A2 cos t sin t E AB sin t cos t E AB
sin t sin t E B 2
= cos t cos t k sin t sin t k
E ( AB) 0 & E ( A2 ) E ( B 2 ) k ( say)
= k cos t cos t sin t sin t
= kcos t t = kcos a function of time difference. X(t) is wide sense stationary.
6. If the process X(t)=P+Qt, where P and Q are independent random variables with
E(P)= p, E(Q) =q, Var(P) = 12 and Var(Q) = 22 then find E{(X(t)} and R(t1,t2). Is the process
{X(t)} stationary? (APRIL 2017)
Solution:
E{X(t)}=E{P+Qt}=E(P)+E(Q)t=p+qt which is not a constant.
R(t1,t2) =E{X(t1),X(t2)} = E{(p+qt1)(p+qt2)}
=E{p2}+E{PQ}( t1+t2)+E(Q2)t1t2
= 12+p2+pq(t1+t2)+( 22+q2) t1t2. ( where E(PQ)=E(P).E(Q)=pq)
Which is not a functions of t1-t2.
Therefore {X(t)} is not stationary.
at n 1
, n 1,2,3,
1 at n 1
PX (t ) n .
at
1 at , n 0.
1 at at 2
1 2 3
1 at 2 1 at 3 1 at 4
2
1 at at
2
1 at
2
1 2 3 1 1 at
1 at 1 at 1 at 1 at
2
2
1 1
1 at 1 a constant
1 at
2
E X 2 t n 2 P X t n = n n 1 n P X t n
n 0 n 0
= n n 1 P X t n nP X t n
n 0
= n n 1 P X t n 1
n 0
nP X t n 1
n 0
0 1.2P X t 1 2.3P X t 2 3.4 P X t 3 1
1 at at 2
2 6 12 1
1 at
2
1 at
3
1 at
4
3
2 at at
2
2 at
1 3 4 1 1 1 at 1
1 at 1 at 1 at 1 at
2 2
3
1 1
1 1 at 1 at , not a constant.
1 at 1 at
2
Solution:
Since y is uniformly distributed in 0, 2 ,
1
f y , 0 y 2
2
2
E X t X t f y dy
0
2
1
2 sin t y dy
0
1 2
cos t y 0
2
1
cos t 2 cost 0
2
RXX t1 , t2 E sin t1 y sin t2 y
cos t1 t2 cos t1 t2 2 y
E
2
E cos t1 t2 E cos t1 t2 2 y
1 1
2 2
2
cos t1 t2 cos t1 t2 2 y
1 1 1
dy
2 2 0 2
1 sin t1 t2 2 y
2
cos t1 t2
1
2 4 2 0
cos t1 t2 sin t1 t2 2 sin t1 t2
1 1
2 8
cos t1 t2 is a function of time difference.
1
2
X t is WSS.
9.Verify whether the sine wave random process X t Y sin t , Y is uniformly distributed in the interval
1,1 is WSS or not
Solution:
Since y is uniformly distributed in 1,1 ,
1
f y , 1 y 1
2
1
E X (t ) X (t ) f ( y )dy
1
1
1
y sint 2 dy
1
sint
1
2 1
ydy
Sint
0 0
2
RXX t1 , t2 E X t1 X t2
E y 2 sint1sint2
cos t1 t2 cosw t1 t2
E y2
2
cos t1 t2 cosw t1 t2
E y2
2
cos t1 t2 cosw t1 t2 1 2
2 1 y f y dy
cos t1 t2 cosw t1 t2 1 1 2
2 1
y dy
2
cos t1 t2 cosw t1 t2 y 3
1
4 3 1
cos t1 t2 cosw t1 t2 2
4 3
cos t1 t2 cosw t1 t2
6
RXX t1 , t2 a function of time difference alone.
Hence it is not a WSS Process.
10.If X t Ycost Zsint for all t & where Y & Z are independent binary random variables. Each of
m
i.e., di GCD m : pij 0
State i is said to be periodic with period d i if di 1 and a periodic if di 1 .
Example:
0 1
1 0 So states are with period 2.
1 3
4 4 The states are aperiodic as period of each state is 1.
1 1
2 2
Ergodic: A non null persistent and aperiodic state is called ergodic.
Example:
0 0 1 0
0 0 0 1
0 1 0 0 Here all the states are ergodic.
1 1 1 1
2 8 8 4
2. Let the Markov Chain consisting of the states 0, 1, 2, 3 have the transition probability matrix
1 1
0 0
2 2
1 0 0 0
P . Determine which states are transient and which are recurrent by defining
0 1 0 0
0 1 0 0
transient and recurrent states. (MAY 2010)
Solution:
Transient state: A state ‘a’ is transient if Faa< 1.Recurrent state: A state ‘a’ is recurrent if Faa=1.
Here Faa f , where f first time return probability of state 'a' after n steps.
n 0
n
aa
n
aa
0 1 1
1/2
1/2
1
1
2 3
1 1
2 0 2
P2 0 1 0
1 1
0
2 2
0 1 0
1 1
P 3 P 2 .P 0 P
2 2
0 1 0
1 1
2 0 2
P 4 P 2 .P 2 0 1 0 P 2
1 1
0
2 2
P2n P2 & P2 n1 P
Also P002 0, P011 0, P022 0
P101 0, P112 0, P121 0
P202 0, P211 0, P222 0
The Markov chain is irreducible
Also Pii2 Pii4 ... 0 for all i
The states of the chain have period 2. Since the chain is finite irreducible, all states are non null persistent. All
states are not ergodic.
5.A gambler has Rs. 2. He bets Re. 1 at a time and wins Re. 1 with probability ½. He stops playing
if he loses Rs. 2 or wins Rs. 4. i) What is the tpm of the related Markov chain? ii) What is the
probability that he has lost his money at the end of 5 plays?
(MAY 2013)
Solution:
Let Xn denote the amount with the player at the end of the nth round of the play.
The possible values of Xn = State space = {0, 1, 2, 3, 4, 5, 6}
Initial probability distribution= P(0) 0 0 1 0 0 0 0
0 1 2 3 4 5 6
0 1 0 0 0 0 0 0
1 1 1
0 0 0 0 0
2 2
1 1
2 0 0 0 0 0
2 2
(i) The transition probability matrix is = 1 1
3 0 0 0 0 0
2 2
1 1
4 0 0 0 0 0
2 2
1 1
0 0 0 0 0
5 2 2
6 0 0 0 0 0 0 1
(ii) Probability that he has lost his money at the end of 5 plays = P X 5 0
To find this we need P
5
1 0 0 0 0 0 0
1 1
0 0 0 0 0
2 2
1 1
0 0 0 0 0
2 2
1 0 1 1
P P P 0 0 1 0 0 0 0 0
1 1
0 0 0 0 0 0 0 0 0
2 2 2 2
1 1
0 0 0
2
0
2
0
0 0 0 0
1
0
1
2 2
0 0 0 0 0 0 1
1 0 0 0 0 0 0
1 1
0 0 0 0 0
2 2
1 1
0 0 0 0 0
2 2
1
P P P 0
1 1 1 1 1 1
0
2 1
0 0 0 0 0 0 0 0 0 0 0 0
2 2 2 2 4 2 4
1 1
0 0 0
2
0
2
0
0 0 0 0
1
0
1
2 2
0 0 0 0 0 0 1
1 0 0 0 0 0 0
1 1
0 0 0 0 0
2 2
1 1
0 0 0 0 0
2 2
1 1
P P P
1 1 1 1 1 3 1
0
3 2
0 0 0 0 0 0 0 0 0 0 0
4 2 4 2 2 4 4 8 8
1 1
0 0 0
2
0
2
0
0 0 0 0
1
0
1
2 2
0 0 0 0 0 0 1
1 0 0 0 0 0 0
1 1
0 0 0 0 0
2 2
1 1
0 0 0 0 0
2 2
1 3 1
P P P
1 3 1 1 1 5 1
0
4 3
0 0
16
0 0 0 0 0 0 0 0
4 4 8 8 2 2 8 16 4
1 1
0 0 0
2
0
2
0
0 0 0 0
1
0
1
2 2
0 0 0 0 0 0 1
1 0 0 0 0 0 0
1 1
0 0 0 0 0
2 2
1 1
0 0 0 0 0
2 2
3 1 3 1
P P P
5 1 1 1 5 9 1
0
5 4
0
16 16
0 0 0 0 0 0 0 0
8 16 4 2 2 8 32 32 8
1 1
0 0 0
2
0
2
0
0 0 0 0
1
0
1
2 2
0 0 0 0 0 0 1
P X 5 0 3 / 8
6. A raining process is considered as a two state Markov chain. If it rains the state is 0 and if it does not
0.6 0.4
rain the state is 1. The TPM is P . If the Initial distribution is 0.4, 0.6 . Find it chance that it
0.2 0.8
will rain on third day assuming that it is raining today.
Solution:
0.6 0.4 0.6 0.4
P2
0.2 0.8 0.2 0.8
0.44 0.56
0.28 0.32
P[rains on third day / it rains today P X 3 0 / X1 0
P002 0.44
7.The transition probability matrix of a Markov chain { X(t)}, n = 1,2,3,….} having three states 1, 2
0.1 0.5 0.4
0.2 0.2 and the initial distribution is P 0 0.7 0.2 0.1 .Find
and 3 is P 0.6
0.3 0.4 0.3
(i) PX 2 3 (ii ) PX 3 2, X 2 3, X 1 3, X 0 2 .
(MAY 2012,2014, DEC 2013)
Solution:
0.43 0.31 0.26
We have P P.P 0.24 0.42 0.34
2
P X 2 3 / X 0 1 P X 0 1 P X 2 3 / X 0 2 P X 0 2 P X 2 3 / X 0 3 P X 0 3
P132 P X 0 1 P232 P X 0 2 P332 P X 0 3 0.26 0.7 0.34 0.2 0.29 0.1 0.279
(II) P X 3 2, X 2 3, X1 3, X 0 2
P X 0 2, X1 3, X 2 3 P X 3 2 / X 0 2, X1 3, X 2 3
P X 0 2, X1 3, X 2 3 P X 3 2 / X 2 3
P X 0 2, X1 3 P X 2 3 / X 0 2, X1 3 P X 3 2 / X 2 3
P X 0 2, X1 3 P X 2 3 / X1 3 P X 3 2 / X 2 3
P X 0 2 P X1 3 / X 0 2 P X 2 3 / X1 3 P X 3 2 / X 2 3
0.2 0.2 0.3 0.4 0.0048
8.Let X n ; n 1, 2,3,..... be a Markov chain with state space S 0,1, 2 and 1 – step Transition probability
0 1 0
1 1 1
matrix P (i) Is the chain ergodic? Explain (ii) Find the invariant probabilities.
4 2 4
0 1 0
Solution:
1 1 1
0 1 0 0 1 0 4 2 4
1
1 1 1 1 1 1 3 1
P 2 P.P
4 2 4 4 2 4 8 4 8
0
1 0 0 1 0 1 1 1
4 2 4
1 1 1 1 3 1
4 2 4 0 1 0 8 4 8
1 3 1 1 1 1 3 5 3
P P P
3 2
8 4 8 4 2 4 16 8 16
1 1 1 0 1 0 1 3 1
4 2 4 8 4 8
P11 0, P13 0, P21 0, P22 0, P33 0 and all other Pij 0
(3) (2) (2) (2) (2) (1)
Therefore the chain is irreducible as the states are periodic with period 1
i.e., aperiodic since the chain is finite and irreducible, all are non null persistent
The states are ergodic.
0 1 0
1 1 1
0 1 2 0 1 2
4 2 4
0 1 0
1
0 (1)
4
1
0 2 1 (2)
2
1
2 (3)
4
0 1 2 1 (4)
From (2) 0 2 1 1 1
2 2
0 1 2 1
1
1 1
2
3 1
1
2
2
1
3
1
From (3) 2
4
1
2
6
2 1
Using (4) 0 1
3 6
4 1
0 1
6
5 1
0 1 0
6 6
1 2 1
0 , 1 & 2 .
6 3 6
9. A man either drives a car or catches a train to go to office each day. He never goes 2 days in a row by train
but he drives one day, then the next day he is just as likely to drive again as he is to travel by train. Now
suppose that one the first day of the week, the man tossed a fair dice and drove to work iff a 6 appeared. Find
the probability that he takes a train on the third day and also the probability that on the third day and also
the probability that he drives to work in the long run.
Solution:
State Space = (train, car)
The TPM of the chain is
T C
PT 0 1
1 1
C
2 2
1
P (traveling by car) = P (getting 6 in the toss of the die)=
6
5
& P (traveling by train) =
6
5 1
P (1) ,
6 6
0 1
5 1 1 11
P P P , 1 1 ,
(2) (1)
6 6 12 12
2 2
0 1
1 11 11 13
P P P , 1 1 ,
3 2
12 12 24 24
2 2
11
P (the man travels by train on the third day) =
24
Let 1 , 2 be the limiting form of the state probability distribution or stationary state distribution of the
Markov chain.
By the property of , P
0 1
1 2 1 1 1 2
2 2
1
2 1
2
1
1 2 2
2
& 1 2 1
1 2
Solving 1 & 2
3 3
2
P{The man travels by car in the long run }= .
3
10.Three are 2 white marbles in urn A and 3 red marbles in urn B. At each step of the process, a marble is
selected from each urn and the 2 marbles selected are inter changed. Let the state ai of the system be the
number of red marbles in A after i changes. What is the probability that there are 2 red marbles in A after 3
steps? In the long run, what is the probability that there are 2 red marbles in urn A?
Solution:
State Space X n 0,1, 2 Since the number of ball in the urn A is always 2.
0 1 2
0 1 0
0
P 1 1 1
1
6 2 3
2
0 2 1
3 3
X n 0 , A 2W (Marbles) B 3R (Marbles)
X n1 0 P00 0
X n1 1 P01 1
X n1 2 P02 0
X n 0 , A 1W &1R (Marbles) B 2R &1W (Marbles)
1
X n 1 0 P10
6
1
X n 1 1 P11
2
1
X n 1 2 P12
3
X n 2 , A 2R (Marbles) B 1R & 2W (Marbles)
X n1 0 P20 0
2
X n 1 1 P21
3
1
X n 1 2 P22
3
P 1, 0, 0 as there is not red marble in A in the beginning.
0
P P P 0, 1, 0
1 0
1 1 1
P P P , ,
2 1
6 2 3
1 23 5
P P P ,
3 2
,
12 36 18
5
P (There are 2 red marbles in A after 3 steps) P X 3 2 P23
18
Let the stationary probability distribution of the chain be 0 , 1 , 2 .
By the property of , P & 0 1 2 1
0 1 0
0 1 2
1 1 1
0 1 2
6 2 3
2 1
0
3 3
1
1 0
6
1 2
0 1 2 1
2 3
1 1
1 2 2
3 3
& 0 1 2 1
1 6 3
Solving 0 , 1 , 2
10 10 10
P {here are 2 red marbles in A in the long run} = 0.3.
11.A salesman territory consists of three cities A, B and C. He never sells in the same city on
successive days. If he sells in city A, then the next day he sells in city B. However if he sells in either
city B or city C, the next day he is twice as likely to sell in city A as in the other city.In the long run
how often does he sell in each of the cities?
(MAY 2012,DEC 2013)
Solution:
States: A, B and C
0 1 0
The transition probability matrix is given by P 2 / 3 0 1/ 3
2 / 3 1/ 3 0
The long run probability is given by a b c , where P = .
0 1 0
Now P = a b c 2 / 3 0 1/ 3 a b c
2 / 3 1/ 3 0
2 2 2 2
0.a + b + c = a -a + b + c = 0 (1)
3 3 3 3
1 1
1.a + 0.b + c b a - b + c 0 (2)
3 3
1 1
0.a + b + 0.c = c b - c = 0 (3)
3 3
Also, we know that a + b + c = 1 (4)
From (3), c = b/3
From (2), a - b + b/9 = 0 a = 8b/9
From (4), 8b/9 + b + b/3 = 1 20b/9 = 1 b = 9/20
c = 3/20 and a = 8/20
Long run probability = 8 / 20 9 / 20 3 / 20
12.The transition probability matrix of a Markov chain { X(t)}, n = 1,2,3,….} having three states 1, 2
0.1 0.5 0.4
and 3 is P 0.6 0.2 0.2 and the initial distribution is P 0 0.7 0.2 0.1 .Find
0.3 0.4 0.3
(i) PX 2 3 (ii ) PX 3 2, X 2 3, X 1 3, X 0 2 .
(MAY 2012,2014, DEC 2013)
Solution:
0.43 0.31 0.26
We have P P.P 0.24 0.42 0.34
2
P X 2 3 / X 0 1 P X 0 1 P X 2 3 / X 0 2 P X 0 2 P X 2 3 / X 0 3 P X 0 3
P132 P X 0 1 P232 P X 0 2 P332 P X 0 3 0.26 0.7 0.34 0.2 0.29 0.1 0.279
(II) P X 3 2, X 2 3, X1 3, X 0 2
P X 0 2, X1 3, X 2 3 P X 3 2 / X 0 2, X1 3, X 2 3
P X 0 2, X1 3, X 2 3 P X 3 2 / X 2 3
P X 0 2, X1 3 P X 2 3 / X 0 2, X1 3 P X 3 2 / X 2 3
P X 0 2, X1 3 P X 2 3 / X1 3 P X 3 2 / X 2 3
P X 0 2 P X1 3 / X 0 2 P X 2 3 / X1 3 P X 3 2 / X 2 3
0.2 0.2 0.3 0.4 0.0048
0 1
13.If the transition probability matrix of a markov chain is 1
1 find the steady-state distribution of the chain.
2 2
Solution:
Let 1 , 2 be the limiting form of the state probability distribution on stationary state distribution of
the markov chain.
By the property of , P
0 1
i.e., 1 , 2 1 1 1 , 2
2 2
1
2 1 ----------- (1)
2
1
1 2 2 ------ (2)
2
Equation (1) & (2) are one and the same.
Consider (1) or (2) with 1 2 1 , since is a probability distribution.
1 2 1
1
Using (1) , 2 2 1
2
3 2
1
2
2
2
3
2 1
1 1 2 1
3 3
1 2
2 1 1 1
3 3
1 2
1 & 2 .
3 3
(v) The probability that the event occurs in a specified number of times t0 , t0 t depends only on t , but not on t0
.
et t
x
e ux
x 0 x!
eu t 1 e2u t 2
e t 1 ...
1! 2!
et ete
u
t eu 1
M X t u e
Let X 1 t and X 2 t be two independent Poisson processes
Their moment generating functions are,
teu 1 2teu 1
M X1t u e 1 and M X 2 t u e
M X1t X 2 t u M X1t u M X 2 t u
teu 1
t eu 1
e 1 e2
teu 1
e 1 2
By uniqueness of moment generating function, the process X1 t X 2 t is a Poisson process with
occurrence rate 1 2 per unit time.
Prove that (i) difference of two independent Poisson processes is not a Poisson process and (ii)
Poisson process is a Markov process. (MAY 2013).
(i) Let X t X1 t X 2 t where X1(t) and X2(t) are poisson processes with 1 and 2 as the parameters
E X t E X1 t E X 2 t 1 2 t
E X 2 t E X1 t X 2 t
2
E X 1
2
t E X 22 t 2E X1 t X 2 t
E X12 t E X 2 2 t 2E X1 t E X 2 t
12t 2 1t 22t 2 2t 2 1t 2t 1 2 t 12 2 2 t 2 212t 2
1 2 t 1 2 t 2 1 2 t 1 2 t 2
2 2
e t3 n3 t1n1 t2 t1 2 1 t3 t2 3
n n n n2
t3 t2
n n t3 t2
n3 n2
n1 ! n2 n1 ! n3 n2 ! e 3 2
P X t3 n3 / X t2 n2
e t2 n2 t1n1 t2 t1
n2 n1
n3 n2 !
n1 ! n2 n1 !
P X t3 n3 / X t2 n2 , X t1 n1 P X t3 n3 / X t2 n2
This means that the conditional probability distribution of X t3 given all the past values X t1 n1 , X t2 n2
depends only on the most recent values X t2 n2 .
i.e., The Poisson process possesses Markov property.Consider
P X t1 n1 , X t2 n2 , X t3 n3
P X t3 n3 / X t2 n2 , X t1 n1
P X t1 n1 , X t2 n2
e t3 n3 t1n1 t2 t1 2 1 t3 t2 3
n n n n2
n1 ! n2 n1 ! n3 n2 !
e t2 n2 t1n1 t2 t1
n2 n1
n1 ! n2 n1 !
t3 t2
n n t3 t2
n3 n2
e 3 2
n3 n2 !
P X t3 n3 / X t2 n2 , X t1 n1 P X t3 n3 / X t2 n2
This means that the conditional probability distribution of X t3 given all the past values X t1 n1 , X t2 n2
depends only on the most recent values X t2 n2 .
i.e., The Poisson process possesses Markov property.
(iii). The inter arrival time of a poisson process i.e., with the interval between two successive occurrences of a
1
poisson process with parameter has an exponential distribution with mean .
Proof:
Let two consecutive occurrences of the event be Ei & Ei1 .
Let Ei take place at time instant ti and T be the interval between the occurrences of Ei Ei1 .
Thus T is a continuous random variable.
P T t P Interval between occurrenceof Ei and Ei 1 exceeds t
P Ei 1 does not occur uptotheinstant ti t
P Noevent occurs intheinterval ti , ti t
P X t 0 P0 t
e t
The cumulative distribution function of T is given by
F t P T t 1 et
The probability density function is given by
f t e t , t 0
1
Which is an exponential distribution with mean .
When is a poisson process said to be homogenous?
The rate of occurrence of the event is a constant, then the process is called a homogenous poisson process
Problems under poisson process
1. If the process N t : t 0 is a Poisson process with parameter , obtain P N t n and E N t .
Solution:
Let be the number of occurrences of the event in unit time.
Let Pn t represent the probability of n occurrences of the event in the interval 0,t . i.e., Pn t P X t n
Pn t t P X t t n
P n occurences in the time 0, t t
n occurences in the interval 0, t and no occurences in t , t t or
P n 1occurences in the interval 0, t and 1 occurences in t , t t or
n 2 occurences in the interval 0, t and 2 occurences in t , t t or...
Pn t 1 t Pn1 t t 0 ...
Pn t t Pn t
Pn1 t Pn t
t
Taking the limits as t 0
Pn t Pn 1 t Pn t -------- (1)
d
dt
This is a linear differential equation.
t
Pn t et Pn 1 t et ------------- (2)
0
Lt P0 t t P0 t
P0 t
t 0 t
dP0 t
P0 t
dt
dP0 t
dt
P0 t
log P0 t t c
P0 t et c
P0 t et ec
P0 t et A ----------------- (4)
Putting t 0 we get
P0 0 e0 A A
i.e., A 1
(4) we have
P0 t et
substituting in (3) we get
t
dt t
0
P1 t e t
t
P2 t e P1 t et dt
t
0
t
ettet dt
0
t 2
2
2
et t
2
P2 t e
t
2!
Proceeding similarly we have in general
e t t
n
Pn t P X t n , n 0,1,...
n!
Thus the probability distribution of X t is the Poisson distribution with parameter t .
E X t t .
2. Find the mean and autocorrelation and auto covariance of the Poisson process.
Solution:
The probability law of the poisson process X t is the same as that of a poisson distribution with parameter t
E X t nP X t n
n 0
e t t
n
n
n 0 n!
t t
n 1
e t
n 1!
n 1
t
te et
E X t t
E X 2 t n2 P X t n
n 0
e t t
n
n n n 2
n 0 n!
e t t e t t
n n
n n 1 n
n 0 n! n 0 n!
e t t
n
n n 1 t
n 0 n!
t t t t
2 3 4
e ... t
1 1! 2!
t e e t
2 t t
t t
2
E X 2 t t 2t 2
Var X t t
RXX t1 , t2 E X t1 X t2
E X t1 X t2 X t1 X t1
E X t1 X t2 X t1 E X 2 t1
E X t1 E X t2 X t1 E X 2 t1
Since X t is a process of independent increments.
t1 t2 t1 t1 t12 if t2 t1 (by (1))
2t1t2 t1 if t2 t1
RXX t1, t2 2 t1t2 min t1, t2
Auto Covariance
CXX t1 , t2 RXX t1 , t2 E X t1 E X t2
2t1 , t2 t1 2t1t2
t1 , if t2 t1
min t1 , t2
3. If the customers arrive at a bank according to a poisson process with a mean rate of 2 per minute, find the
probability that, during an 1- minute interval no customer arrives.
Solution:
Here 2 , t 1
e t t
n
P X t n , n 0,1, 2,...
n!
Probability during 1-min interval, no customer arrives P X t 0 e2 .
4. Examine whether the Poisson process {X(t)} is stationary or not. (DEC2010, 2012, APR 2015)
A random process to be stationary in any sense, its mean must be a constant. We know that the mean of a
Poisson process with rate is given by E{X(t)} = t which depends on the time t. Thus the Poisson
process is not a stationary process.
5. Is a Poisson process a continuous time Markov chain? Justify your answer (MAY2010)
We know that Poisson process has the Markovian property. Therefore, it is a Markov chain as the states of
Poisson process are discrete. Also, the time‘t’ in a Poisson process is continuous.Therefore, the Poisson
process a continuous time Markov chain.
6. A radioactive source emits particles at a rate of 5 per min in accordance with Poisson process.
Each particle emitted has a probability 0.6 of being recorded. Find the probability that 10 particles
are recorded in 4 min period. (DEC 2014)
The number of particles N(t) emitted is Poisson with parameter = np = 5(0.6) = 3
e3t 3t
m
P( N (t ) m)
m!
e3(4) 3(4)
10
e3 36
(i) P X 6
1
0.0504
2 6
e12 (12)4
(ii) P X (2) 4 0.053
4
3 3 0
e 2 3
1 1 2
P X 1 1 P X 01 1 e 2 0.776.
(iii) 4
4 0
8.Suppose that customers arrive at a bank according to Poisson process with mean rate of 3 per
minute. Find the probability that during a time interval of two minutesi) exactly 4 customers arrive
ii) greater than 4 customers arrive iii) fewer than 4 customers arrive.
(MAY 2012, DEC 2013)
Solution:
Mean of the Poisson process = t.
Mean arrival rate = mean number of arrivals per minute (unit time) =
et t
k
Given = 3. P X t k
k
e6 64
(i) Probability that exactly 4 customers arrive P X 2 4 0.133
4
(ii) Probability that greater than 4 customers arrive
P X 2 4 1 P X 2 0 P X 2 1 P X 2 2 P X 2 3 P X 2 4
e6 60 e6 61 e6 62 e6 63 e6 64
1 0.715
0 1 2 3 4
(iii) Probability that fewer than 4 customers arrive
P X 2 4 P X 2 0 P X 2 1 P X 2 2 P X 2 3
e6 60 e6 61 e6 62 e6 63
0.151
0 1 2 3