Random Process and Linear Algebra: Unit III: Random Processes,,

Discrete Parameter Markov Process [Markov Chain]

Chapman Kolomogorov equations-Limiting distributions

A discrete parameter Markov process is called a Markov Chain. example : The simple random walk moves, at each step, to a randomly chosen nearest neighbour.

Discrete Parameter Markov Process [Markov Chain]

- Chapman Kolmogorov equations - Limiting distributions

A discrete parameter Markov process is called a Markov Chain.

Note: Mathematically, we define the Markov Chain as follows


then the process Xn; n = 0, 1, 2, ... are called as Markov Chain. Here a1, a2, ..., an are called the states of the Markov chain.

Example :

The simple random walk moves, at each step, to a randomly chosen nearest neighbour.

Definition: One-step Transition Probability

The conditional probability P {Xn = aj | Xn-1 = ai} is called the one-step transition probability from state ai to state aj at the nth step (trial) and is denoted by Pij (n-1, n).

Definition : Homogeneous Markov Chain

If Pij (n-1, n) = Pij (m-1, m), then the Markov Chain is called the homogeneous Markov Chain or the Chain is said to have stationary transition probabilities.

Note: The use of the word 'Stationary' does not imply a stationary random sequence.

Definition : Transition Probability Matrix (TPM)

When the Markov Chain is homogeneous, the one step transition probability is denoted by Pij. The matrix P = {Pij} is called (one-step) transition probability matrix (tpm).

Note: The tpm of a chain is a Stochastic matrix, since all Pij ≥ 0 and the sum of all elements of any row of the transition matrix is equal to 1.

Definition: n-Step Transition Probability [A.U A/M 2019 R13 PQT]

The conditional probability that the process is state aj at step n, given that it was in state ai at step 0. (i.e.,) P{Xn = aj / Xn = ai} is called the n-step transition probability and denoted by 

Note :


Definition: Chapman - Kolmogorov Theorem :

If 'P' is the tpm of a homogeneous Markov chain, then the n-step tpm p(n) is equal to pn.


Derive the Chapman-Kolmogorov equations for discrete-time Markov chain.

[A.U. N/D 2007, A.U A/M 2015]


Definition: Regular matrix

A stochastic matrix 'P' is said to be a regular matrix, if all the entries of pm (for some +ve integer m) are +ve. A homogeneous Markov chain is said to be regular if its tpm is regular.

We state below two theorems without proofs.

1. If P = {Pij} is the state probability distribution of the process at an arbitrary time, then that after one step is P, where  is the tpm of the chain and that after 'n' steps is P

2. If a homogeneous Markov chain is regular, then every sequence of state probability distributions approaches a unique fixed probability distribution called the stationary (state) distribution or steady-state distribution of the Markov chain.

 where the state probability distribution at step n,


3. Moreover, if 'P' is the tpm of the regular chain, then лP = π (π is a row vector). Using this property of л, it can be found out.


Hint :

Steady state (or) invariant (or) stationary (or) limiting-state (or)  (or) long run (or) thousandth trial these are represents the same л


Note :-

(1) When all the entries of a tpm are positive,

i.e., if the tpm is regular and of the form


(2) If the tpm of a chain is a stochastic matrix, then the sum of all elements of any row is equal to 1.

Example 3.7.1

If the tpm of a Markov chain is  find the steady statedaistribution of the chain. [A.U N/D 2019 R17 PQT] [A.U A/M 2019 R17 PQT] 51832-gnitimil [AU CBT Dec. 2009, Trichy M/J 2011, Tvli M/J 2011] [A.U M/J 2013] [A.U A/M 2017 R-8] [A.U A/M 2018 R-13]

Solution:

Given: 

Let л = (л1, л2) is the steady state distribution of the chain.



Aliter : If π = (π1, π2) is the steady state distribution of the chain, then by the property of л, we have



Example 3.7.2

The three-state Markov chain is given by the tpm  Find the steady-state distribution of the chain.

Solution:

If π = (π1 π2 π3) is the steady-state distribution of the chain, then by the property of π, we have






.'. The steady state distribution of the chain is


Example 3.7.3

An Engineer analyzing a series of digital signals generated by a testing system observes that only 1 out of 15 highly distorted signals followed a highly distorted signal with no recognizable signal, whereas 20 out of 23 recognizable signals follow recognizable signals with no highly distorted signals between. Given that only highly distorted signals are not recognizable. Find the fraction of signals that are highly distorted. [AU Dec. 2004, N/D 2007, M/J 2009, N/D 2010, N/D 2014] [A.U A/M 2015]

Solution :

Let π1 -> the fraction of signals that are recognizable. [R]

π2 -> the fraction of signals that are highly distorted. [D]

The tpm of the Markov chain is




Example 3.7.4.

A student's study habits are as follows: If he studies one night, he is 70% sure not to study next night. On the otherhand, if he does not study one night, he is 60% sure not to study the next night as well. In the long run, how often does he study?

Solution :

Let π1 -> he studies one night [S]

π2 -> he does not study one night [N]

The tpm of the Markov Chain is


Let π = (π1, π2) is the steady state distribution of the chain.


Example 3.7.5.

Two boys B1, B2 and two girls G1, G2 are throwing a ball from one to another. Each boy throws the ball to the other boy with probability 1/2 and to each girl with probability 1/4. On the other hand, each girl throws the ball to each boy with probability 1/2 and never to the other girl. In the long run, how often does each receive the ball? [A.U CBT N/D 2011] [A.U N/D 2013 R-08]

Solution :

Let π1 -> boy (B1)

π2 -> boy (B2)

π3 -> girl (G1)

π4 -> girl (G2)

The transition diagram is


.'. The tpm of the Markov Chain is


['.' If the tpm of a chain is a stochastic matrix, then the sum of all elements of any row is equal to 1]

If π = (π1 π2 π3 π4) is the steady state distribution of the chain, then by the property of л, we have

πP = 1 .............(1)

π1 + π2 + π3 + π4 = 1 .............(2)




.'. The steady state distribution of the chain is

π = (π1 π2 π3 π4)

i.e., π = (1/3 1/3 1/6 1/6)

Thus, in the long run, the probability that each boy receives the ball = 1/3, the probability that each girl receives the ball = 1/6

Example 3.7.6.

A salesman's territory consists of three cities A, B and C. He never sells in the same city on successive days. If he sells in A, then the next day he sells in city B. However, if he sells in either B or C, then the next day he is twice as likely to sell in city A as in the other city. In the long run, how often does he sell in each of the cities? [A.U M/J 2013] [A.U N/D 2013, M/J 2012] [A.U A/M 2017 R-8, N/D 2017 R-8]

Solution :

Let π1 -> City A

π2 -> City B

π3 -> City C

.'. The tpm of the Markov chain is




If π = (π1 π2 π3) is the steady state distribution of the chain, then by the property of π, we have


In equation (2), we change π1 and π3 interms of π2




.'. The steady-state distribution of the chain is π= (π1 π2 π3)


Thus in the long run, he sells 40% of the item in city A, 45% of the item in city B and 15% of the item in city C.

Example 3.7.7.

A housewife buys 3 kinds of cereals A, B and C. She never buys the same cereal in successive weeks. If she buys cereal A, the next week she buys cereal B. However if she buys B or C, the next week she is 3 times as likely to buy A as the other cereal. In the long run how often she buys each of the 3 cereals? [CBT A/M 2011, Trichy M/J 2011] [AU Tvli. M/J 2010]

Solution:

Given: Let π1 → Cereal А; π2 → Cereal В; π3 → Cereal C

.'. The tpm of the Markov chain is



If π = (π1 π2 π3) is the steady-state distribution of the chain, then by the property of π, we have




Example 3.7.8.

A man is at an integral part on the X-axis between the origin and the point 3. He takes a unit step to the right with probability or 1/3 to the left with probability 2/3, unless he is at the origin, where he of takes a step to the right to reach the point 1 or is at the point 3, where he takes a step to the left to reach the point 2. What is the probability that he is at the point 1 in the long run?

Solution:

The transition diagram is


Given:

Let π1 → Integral part 0, π2 → Integral part 1,

π3 → Integral part 2, π4 → Integral part 3

.'. The tpm of the Markov chain




If π = (π1 π2 π3 π4) is the steady-state distribution of the chain, then by the property of π, we have

πP = π .............(1)

π1 + π2 + π3 + π4 = 1 .............(2)




The steady-state distribution of the chain is

π = (π1 π2 π3 π4)

i.e., π = (2/7 3/7 3/14 1/14)

.'. The probability that he is at 1 after a long run = 3/7

Random Process and Linear Algebra: Unit III: Random Processes,, : Tag: : Chapman Kolomogorov equations-Limiting distributions - Discrete Parameter Markov Process [Markov Chain]