SlideShare a Scribd company logo
Chapter 17 Markov Chains
Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves over time includes stochastic processes. An explanation of stochastic processes – in particular, a type of stochastic process known as a Markov chain is included. We begin by defining the concept of a stochastic process.
5.1 What is a Stochastic Process? Suppose we observe some characteristic of a system at discrete points in time. Let  X t  be the value of the system characteristic at time  t . In most situations,  X t  is not known with certainty before time  t  and may be viewed as a random variable. A  discrete-time stochastic process  is simply a description of the relation between the random variables  X 0,  X 1,  X 2  ….. Example: Observing the price of a share of Intel at the beginning of each day Application areas: education, marketing, health services, finance, accounting, and production
A  continuous –time stochastic process  is simply the stochastic process in which the state of the system can be viewed at any time, not just at discrete instants in time. For example, the number of people in a supermarket  t  minutes after the store opens for business may be viewed as a continuous-time stochastic process.
5.2 What is a Markov Chain? One special type of discrete-time is called a Markov Chain. Definition:  A discrete-time stochastic process is a  Markov chain  if, for  t  = 0,1,2… and all states P ( X t+1  =  i t+1 | X t  = i t ,  X t-1 = i t-1 ,…, X 1 = i 1 ,  X 0 = i 0 )   = P ( X t+1 = i t+1 | X t  =  i t ) Essentially this says that the probability distribution of the state at time  t +1 depends on the state at time  t ( i t ) and does not depend on the states the chain passed through on the way to  i t  at time  t .
In our study of Markov chains, we make further assumption that for all states  i  and  j  and all  t ,  P ( X t+1  =  j | X t  =  i ) is independent of  t . This assumption allows us to write  P ( X t+1  =  j | X t  =  i ) =  p ij  where  p ij   is the probability that given the system is in state  i  at time  t , it will be in a state  j  at time  t +1. If the system moves from state  i  during one period to state  j  during the next period, we call that a  transition  from  i  to  j  has occurred.
The  p ij ’s are often referred to as the  transition probabilities  for the Markov chain. This equation implies that the probability law relating the next period’s state to the current state does not change over time. It is often called the  Stationary Assumption  and any Markov chain that satisfies it is called a  stationary Markov chain . We also must define  q i  to be the probability that the chain is in state  i  at the time 0; in other words,  P ( X 0 = i ) =  q i .
We call the vector  q = [ q 1 , q 2 ,…q s ] the  initial probability distribution  for the Markov chain. In most applications, the transition probabilities are displayed as an  s  x  s   transition probability matrix   P . The transition probability matrix  P  may be written as
For each  I We also know that each entry in the  P  matrix must be nonnegative. Hence, all entries in the transition probability matrix are nonnegative, and the entries in each row must sum to 1.
The Gambler’s Ruin Problem At time 0, I have $2. At times 1, 2, …, I play a game in which I bet $1, with probabilities p, I win the game, and with probability 1 – p, I lose the game. My goal is to increase my capital to $4, and as soon as I do, the game is over. The game is also over if my capital is reduced to 0. Let X t  represent my capital position after the time t game (if any) is played X 0 , X 1 , X 2 , …. May be viewed as a discrete-time stochastic process
The Gambler’s Ruin Problem $0  $1  $2  $3  $4
5.3  n -Step Transition Probabilities A question of interest when studying a Markov chain is: If a Markov chain is in a state  i  at time  m , what is the probability that  n  periods later the Markov chain will be in state  j ? This probability will be independent of  m , so we may write P ( X m+n  = j | X m  =  i ) =  P ( X n  = j | X 0  =  i ) =  P ij ( n ) where  P ij ( n ) is called the  n -step probability  of a transition from state  i  to state  j . For  n  > 1,  P ij ( n ) =  ij th element of  P n  P ij (2) is the (i, j)th element of matrix P 2  = P 1  P 1 P ij (n) is the (i, j)th element of matrix P n  = P 1  P n-1
The Cola Example Suppose the entire cola industry produces only two colas. Given that a person last purchased cola 1, there is a 90% chance that their next purchase will be cola 1. Given that a person last purchased cola 2, there is an 80% chance that their next purchase will be cola 2. If a person is currently a cola 2 purchaser, what is the probability that they will purchase cola 1 two purchases from now? If a person is currently a cola 1 a purchaser, what is the probability that they will purchase cola 1 three purchases from now?
The Cola Example We view each person’s purchases as a Markov chain with the state at any given time being the type of cola the person last purchased. Hence, each person’s cola purchases may be represented by a two-state Markov chain, where State 1 = person has last purchased cola 1 State 2 = person has last purchased cola 2 If we define  X n  to be the type of cola purchased by a person on her  n th future cola purchase, then  X 0 ,  X 1 , … may be described as the Markov chain with the following transition matrix:
The Cola Example We can now answer questions 1 and 2. We seek  P ( X 2  = 1| X 0  = 2) =  P 21 (2) = element 21 of  P 2 :
The Cola Example Hence,  P 21 (2) =.34. This means that the probability is .34 that two purchases in the future a cola 2 drinker will purchase cola 1. We seek  P 11 (3) = element 11 of  P 3 : Therefore,  P 11 (3) = .781
Many times we do not know the state of the Markov chain at time 0. Then we can determine the probability that the system is in state  i  at time  n  by using the reasoning. Probability of being in state  j  at time  n  where  q =[q 1 , q 2 , … q 3 ]. Hence, q n  = q o p n  = q n-1 p Example, q 0  = (.4,.6) q 1 =  (.4,.6)  q 1  = (.48,.52)
To illustrate the behavior of the  n -step transition probabilities for large values of  n , we have computed several of the  n -step transition probabilities for the Cola example. This means that for large  n,  no matter what the initial state, there is a .67 chance that a person will be a cola 1 purchaser.
5.4 Classification of States in a Markov Chain To understand the  n -step transition in more detail, we need to study how mathematicians classify the states of a Markov chain. The following transition matrix illustrates most of the following definitions. A graphical representation is shown in the book (State-Transition diagram)
Definition:  Given two states of  i  and  j , a  path  from  i  to  j  is a sequence of transitions that begins in  i  and ends in  j , such that each transition in the sequence has a positive probability of occurring.  Definition:  A state  j  is  reachable  from state  i  if there is a path leading from  i  to  j . Definition:  Two states  i  and  j  are said to  communicate  if  j  is reachable from  i , and  i  is reachable from  j . Definition:  A set of states  S  in a Markov chain is a  closed set  if no state outside of  S  is reachable from any state in  S .
Definition:  A state  i  is an  absorbing state  if  p ij =1. Definition:  A state  i  is a  transient state  if there exists a state  j  that is reachable from  i , but the state  i  is not reachable from state  j . Definition: If a state is not transient, it is called a recurrent state. Definition: A state  i  is  periodic  with period k > 1 if k is the smallest number such that all paths leading from state  i  back to state  i  have a length that is a multiple of k. If a recurrent state is not periodic, it is referred to as  aperiodic . If all states in a chain are recurrent, aperiodic, and communicate with each other, the chain is said to be  ergodic . The importance of these concepts will become clear after the next two sections.
5.5 Steady-State Probabilities and Mean First Passage Times Steady-state probabilities are used to describe the long-run behavior of a Markov chain. Theorem 1:  Let  P  be the transition matrix for an  s -state ergodic chain. Then there exists a vector  π  = [ π 1   π 2   …  π s ] such that
Theorem 1 tells us that for any initial state  i ,  The vector  π  = [ π 1   π 2   …  π s ]  is often called the  steady-state distribution , or  equilibrium distribution , for the Markov chain. Hence, they are  independent of the initial probability distribution defined over the states
Transient Analysis & Intuitive Interpretation The behavior of a Markov chain before the steady state is reached is often call  transient  (or short-run)  behavior . An interpretation can be given to the steady-state probability equations. This equation may be viewed as saying that in the steady-state, the “flow” of probability into each state must equal the flow of probability out of each state.
Steady-State Probabilities The vector    = [  1 ,   2 , …. ,   s  ] is often known as the steady-state distribution for the Markov chain For large n and all i, P ij (n+1)    P ij (n)      j In matrix form    =   P For any n and any i, P i1 (n) + P i2 (n) + … + P is (n) = 1 As n  , we have   1  +   2  + …. +   s  = 1
An Intuitive Interpretation of Steady-State Probabilities Consider Subtracting   j p jj  from both sides of the above equation, we have Probability that a particular transition enters state j = probability that a particular transition leaves state j
Use of Steady-State Probabilities in Decision Making In the Cola Example, suppose that each customer makes one purchase of cola during any week.  Suppose there are 100 million cola customers.  One selling unit of cola costs the company $1 to produce and is sold for $2. For $500 million/year, an advertising firm guarantees to decrease from 10% to 5% the fraction of cola 1 customers who switch after a purchase. Should the company that makes cola 1 hire the firm?
At present, a fraction  π 1  = ⅔ of all purchases are cola 1 purchases, since: π 1  = .90 π 1 +.20 π 2 π 2  = .10 π 1 +.80 π 2 and using the following equation by  π 1  +  π 2  = 1 Each purchase of cola 1 earns the company a $1 profit. We can calculate the annual profit as $3,466,666,667 [2/3(100 million)(52 weeks)$1]. The advertising firm is offering to change the  P  matrix to
For  P 1, the steady-state equations become π 1  = .95 π 1 +.20 π 2 π 2  = .05 π 1 +.80 π 2 Replacing the second equation by  π 1  +  π 2  = 1 and solving, we obtain  π 1 =.8 and   π 2  = .2.  Now the cola 1 company’s annual profit will be $3,660,000,000 [.8(100 million)(52 weeks)$1-($500 million)]. Hence, the cola 1 company should hire the ad agency.
Inventory Example A camera store stocks a particular model camera that can be ordered weekly. Let D 1 , D 2 , … represent the demand for this camera (the number of units that would be sold if the inventory is not depleted) during the first week, second week, …, respectively. It is assumed that the D i ’s are independent and identically distributed random variables having a Poisson distribution with a mean of 1. Let X 0  represent the number of cameras on hand at the outset, X 1  the number of cameras on hand at the end of week 1, X 2  the number of cameras on hand at the end of week 2, and so on.  Assume that X 0  = 3.  On Saturday night the store places an order that is delivered in time for the next opening of the store on Monday.  The store using the following order policy: If there are no cameras in stock, 3 cameras are ordered. Otherwise, no order is placed.  Sales are lost when demand exceeds the inventory on hand
Inventory Example X t  is the number of Cameras in stock at the end of week t (as defined earlier), where X t  represents the state of the system at time t Given that X t  = i, X t+1  depends only on D t+1  and X t  (Markovian property) D t  has a Poisson distribution with mean equal to one. This means that P(D t+1  = n) = e -1 1 n /n! for n = 0, 1, … P(D t  = 0 ) = e -1  = 0.368 P(D t  = 1 ) = e -1  = 0.368 P(D t  = 2 ) = (1/2)e -1  = 0.184 P(D t     3 ) = 1 – P(D t     2) = 1 – (.368 + .368 + .184) = 0.08 X t+1  = max(3-D t+1 , 0) if X t  = 0 and X t+1  = max(X t  – D t+1 , 0) if X t     1, for t =  0, 1, 2, ….
Inventory Example: (One-Step) Transition Matrix P 03  = P(D t+1  = 0) = 0.368 P 02  = P(D t+1  = 1) = 0.368 P 01  = P(D t+1  = 2) = 0.184 P 00  = P(D t+1     3) = 0.080
Inventory Example: Transition Diagram 0 1 2 3
Inventory Example: (One-Step) Transition Matrix
Transition Matrix: Two-Step P (2)  = PP
Transition Matrix: Four-Step P (4)  = P (2) P (2)
Transition Matrix: Eight-Step P (8)  = P (4) P (4)
Steady-State Probabilities The steady-state probabilities uniquely satisfy the following steady-state equations  0  =   0 p 00  +   1 p 10  +   2 p 20  +   3 p 30  1  =   0 p 01  +   1 p 11  +   2 p 21  +   3 p 31  2  =   0 p 02  +   1 p 12  +   2 p 22  +   3 p 32  3  =   0 p 03  +   1 p 13  +   2 p 23  +   3 p 33 1 =   0  +   1  +   2  +   3
Steady-State Probabilities: Inventory Example  0  = .080  0  + .632  1  + .264  2 + .080  3  1  = .184  0  + .368  1  + .368  2  + .184  3  2  = .368  0   + .368  2  + .368  3  3  = .368  0   + .368  3 1  =   0  +   1  +   2  +   3  0  = .286,   1  = .285,   2  = .263,   3  = .166 The numbers in each row of matrix P (8)  match the corresponding steady-state probability
Mean First Passage Times For an ergodic chain, let  m ij  = expected number of transitions before we first reach state  j , given that we are currently in state  i; m ij  is called the  mean first passage time  from state  i  to state  j . In the example, we assume we are currently in state  i . Then with probability  p ij ,  it will take one transition to go from state  i  to state  j . For  k  ≠ j , we next go with probability  p ik  to state  k . In this case, it will take an average of 1 +  m kj  transitions to go from  i  and  j .
This reasoning implies By solving the linear equations of the equation above, we find all the mean first passage times. It can be shown that
For the cola example,  π 1 =2/3 and   π 2  = 1/3 Hence,  m 11  = 1.5 and   m 22  = 3 m 12  = 1 + p 11 m 12   = 1 + .9m 12 m 21  = 1 + p 22 m 21  = 1 + .8m 21 Solving these two equations yields,  m 12  = 10 and   m 21  = 5
Solving for Steady-State Probabilities and Mean First Passage Times on the Computer Since we solve steady-state probabilities and mean first passage times by solving a system of linear equations, we may use LINDO to determine them. Simply type in an objective function of 0, and type the equations you need to solve as your constraints.
5.6 Absorbing Chains Many interesting applications of Markov chains involve chains in which some of the states are absorbing and the rest are transient states. This type of chain is called an  absorbing chain . To see why we are interested in absorbing chains we consider the following accounts receivable example.
Accounts Receivable Example The accounts receivable situation of a firm is often modeled as an absorbing Markov chain. Suppose a firm assumes that an account is uncollected if the account is more than three months overdue. Then at the beginning of each month, each account may be classified into one of the following states: State 1  New account State 2  Payment on account is one month overdue State 3  Payment on account is two months overdue State 4  Payment on account is three months overdue State 5  Account has been paid State 6  Account is written off as bad debt
Suppose that past data indicate that the following Markov chain describes how the status of an account changes from one month to the next month: New 1 month 2 months 3 months Paid Bad Debt New  1 month  2 months  3 months  Paid  Bad Debt
To simplify our example, we assume that after three months, a debt is either collected or written off as a bad debt. Once a debt is paid up or written off as a bad debt, the account if closed, and no further transitions occur. Hence, Paid or Bad Debt are absorbing states. Since every account will eventually be paid or written off as a bad debt, New, 1 month, 2 months, and 3 months are transient states.
A typical new account will be absorbed as either a collected debt or a bad debt. What is the probability that a new account will eventually be collected? To answer this questions we must write a transition matrix. We assume  s  –  m  transient states and  m  absorbing states. The transition matrix is written in the form of m columns s-m  rows m  rows s-m columns P =
The transition matrix for this example is Then  s  =6,  m  =2, and  Q  and  R  are as shown. New 1 month 2 months 3 months Paid Bad Debt New  1 month  2 months  3 months  Paid  Bad Debt Q R
What is the probability that a new account will eventually be collected? (.964) What is the probability that a one-month overdue account will eventually become a bad debt?  (.06) If the firm’s sales average $100,000 per month, how much money per year will go uncollected? From answer 1, only 3.6% of all debts are uncollected. Since yearly accounts payable are $1,200,000 on the average, (0.036)(1,200,000) = $43,200 per year will be uncollected.

More Related Content

PPTX
Chap 4 markov chains
PPTX
Markov chain
PDF
Environmental studies
PDF
Quick flask an intro to flask
PPTX
System reliability and types of systems in machine design
PPTX
Stochastic Process and its Applications.
PPTX
peer to peer and client server model
PPTX
Hidden markov model
Chap 4 markov chains
Markov chain
Environmental studies
Quick flask an intro to flask
System reliability and types of systems in machine design
Stochastic Process and its Applications.
peer to peer and client server model
Hidden markov model

What's hot (20)

PDF
Lesson 11: Markov Chains
PPTX
Markov chain-model
PDF
Markov Chain Monte Carlo Methods
PPT
markov chain.ppt
PPT
Markov chains1
ODP
Markov chain and its Application
PPTX
Introduction to Maximum Likelihood Estimator
PDF
Markov decision process
PPTX
Hidden Markov Model
PPTX
Markov chain
PPTX
Markov chain
PPTX
Adversarial search
PPTX
Dynamic programming
PDF
17-markov-chains.pdf
PDF
Monte carlo simulation
PDF
Artificial Intelligence Notes Unit 2
PPTX
An introduction to reinforcement learning
PDF
Feature selection
PPTX
Markov decision process
PDF
Reinforcement learning, Q-Learning
Lesson 11: Markov Chains
Markov chain-model
Markov Chain Monte Carlo Methods
markov chain.ppt
Markov chains1
Markov chain and its Application
Introduction to Maximum Likelihood Estimator
Markov decision process
Hidden Markov Model
Markov chain
Markov chain
Adversarial search
Dynamic programming
17-markov-chains.pdf
Monte carlo simulation
Artificial Intelligence Notes Unit 2
An introduction to reinforcement learning
Feature selection
Markov decision process
Reinforcement learning, Q-Learning
Ad

Viewers also liked (7)

PDF
Berger 2000
PPTX
Hr forecasting techniques
PPT
Markov chain intro
PPT
Actuarial Application of Monte Carlo Simulation
PPT
Session 42_1 Peter Fries-Hansen
PPT
Markov analysis
PPT
construction risk factor analysis: BBN Network
Berger 2000
Hr forecasting techniques
Markov chain intro
Actuarial Application of Monte Carlo Simulation
Session 42_1 Peter Fries-Hansen
Markov analysis
construction risk factor analysis: BBN Network
Ad

Similar to Markov Chains (20)

PDF
Markov chain
PPTX
Stochastic matrices
PPTX
Machine learning fundamental concepts in detail
PDF
12 Machine Learning Supervised Hidden Markov Chains
PPTX
Business Analytics CHAPTER 16 STOCHASTIC MODELS Business Analytics.PPTX
PDF
Book chapter-5
PDF
A STUDY ON MARKOV CHAIN WITH TRANSITION DIAGRAM
PDF
PPT
Hidden Markov Models with applications to speech recognition
PPT
Hidden Markov Models with applications to speech recognition
PPTX
Stochastic Processes Assignment Help
PDF
Martingales Stopping Time Processes
PPTX
Stat 2153 Stochastic Process and Markov chain
PDF
Fundamentos de la cadena de markov - Libro
PDF
By BIRASA FABRICE
PPT
Chapter0
PDF
Hw3 2016 | Applied Stochastic process MTH 412 IIT Kanpur
PDF
ResearchPaper
DOCX
IE 423 page 1 of 1 •••••••••••••••••••••••••••••••••••••••.docx
PDF
Queueing theory
Markov chain
Stochastic matrices
Machine learning fundamental concepts in detail
12 Machine Learning Supervised Hidden Markov Chains
Business Analytics CHAPTER 16 STOCHASTIC MODELS Business Analytics.PPTX
Book chapter-5
A STUDY ON MARKOV CHAIN WITH TRANSITION DIAGRAM
Hidden Markov Models with applications to speech recognition
Hidden Markov Models with applications to speech recognition
Stochastic Processes Assignment Help
Martingales Stopping Time Processes
Stat 2153 Stochastic Process and Markov chain
Fundamentos de la cadena de markov - Libro
By BIRASA FABRICE
Chapter0
Hw3 2016 | Applied Stochastic process MTH 412 IIT Kanpur
ResearchPaper
IE 423 page 1 of 1 •••••••••••••••••••••••••••••••••••••••.docx
Queueing theory

Recently uploaded (20)

PPTX
Probability Distribution, binomial distribution, poisson distribution
PPTX
CkgxkgxydkydyldylydlydyldlyddolydyoyyU2.pptx
PDF
IFRS Notes in your pocket for study all the time
PPTX
ICG2025_ICG 6th steering committee 30-8-24.pptx
PDF
Cours de Système d'information about ERP.pdf
PDF
How to Get Business Funding for Small Business Fast
PPT
340036916-American-Literature-Literary-Period-Overview.ppt
PPTX
Board-Reporting-Package-by-Umbrex-5-23-23.pptx
DOCX
Business Management - unit 1 and 2
PPT
Chapter four Project-Preparation material
PPTX
svnfcksanfskjcsnvvjknsnvsdscnsncxasxa saccacxsax
PDF
Laughter Yoga Basic Learning Workshop Manual
PPT
Lecture 3344;;,,(,(((((((((((((((((((((((
PPTX
3. HISTORICAL PERSPECTIVE UNIIT 3^..pptx
PDF
Hindu Circuler Economy - Model (Concept)
PDF
NISM Series V-A MFD Workbook v December 2024.khhhjtgvwevoypdnew one must use ...
PDF
Nidhal Samdaie CV - International Business Consultant
PDF
Solara Labs: Empowering Health through Innovative Nutraceutical Solutions
PDF
Ôn tập tiếng anh trong kinh doanh nâng cao
PDF
Comments on Crystal Cloud and Energy Star.pdf
Probability Distribution, binomial distribution, poisson distribution
CkgxkgxydkydyldylydlydyldlyddolydyoyyU2.pptx
IFRS Notes in your pocket for study all the time
ICG2025_ICG 6th steering committee 30-8-24.pptx
Cours de Système d'information about ERP.pdf
How to Get Business Funding for Small Business Fast
340036916-American-Literature-Literary-Period-Overview.ppt
Board-Reporting-Package-by-Umbrex-5-23-23.pptx
Business Management - unit 1 and 2
Chapter four Project-Preparation material
svnfcksanfskjcsnvvjknsnvsdscnsncxasxa saccacxsax
Laughter Yoga Basic Learning Workshop Manual
Lecture 3344;;,,(,(((((((((((((((((((((((
3. HISTORICAL PERSPECTIVE UNIIT 3^..pptx
Hindu Circuler Economy - Model (Concept)
NISM Series V-A MFD Workbook v December 2024.khhhjtgvwevoypdnew one must use ...
Nidhal Samdaie CV - International Business Consultant
Solara Labs: Empowering Health through Innovative Nutraceutical Solutions
Ôn tập tiếng anh trong kinh doanh nâng cao
Comments on Crystal Cloud and Energy Star.pdf

Markov Chains

  • 2. Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves over time includes stochastic processes. An explanation of stochastic processes – in particular, a type of stochastic process known as a Markov chain is included. We begin by defining the concept of a stochastic process.
  • 3. 5.1 What is a Stochastic Process? Suppose we observe some characteristic of a system at discrete points in time. Let X t be the value of the system characteristic at time t . In most situations, X t is not known with certainty before time t and may be viewed as a random variable. A discrete-time stochastic process is simply a description of the relation between the random variables X 0, X 1, X 2 ….. Example: Observing the price of a share of Intel at the beginning of each day Application areas: education, marketing, health services, finance, accounting, and production
  • 4. A continuous –time stochastic process is simply the stochastic process in which the state of the system can be viewed at any time, not just at discrete instants in time. For example, the number of people in a supermarket t minutes after the store opens for business may be viewed as a continuous-time stochastic process.
  • 5. 5.2 What is a Markov Chain? One special type of discrete-time is called a Markov Chain. Definition: A discrete-time stochastic process is a Markov chain if, for t = 0,1,2… and all states P ( X t+1 = i t+1 | X t = i t , X t-1 = i t-1 ,…, X 1 = i 1 , X 0 = i 0 ) = P ( X t+1 = i t+1 | X t = i t ) Essentially this says that the probability distribution of the state at time t +1 depends on the state at time t ( i t ) and does not depend on the states the chain passed through on the way to i t at time t .
  • 6. In our study of Markov chains, we make further assumption that for all states i and j and all t , P ( X t+1 = j | X t = i ) is independent of t . This assumption allows us to write P ( X t+1 = j | X t = i ) = p ij where p ij is the probability that given the system is in state i at time t , it will be in a state j at time t +1. If the system moves from state i during one period to state j during the next period, we call that a transition from i to j has occurred.
  • 7. The p ij ’s are often referred to as the transition probabilities for the Markov chain. This equation implies that the probability law relating the next period’s state to the current state does not change over time. It is often called the Stationary Assumption and any Markov chain that satisfies it is called a stationary Markov chain . We also must define q i to be the probability that the chain is in state i at the time 0; in other words, P ( X 0 = i ) = q i .
  • 8. We call the vector q = [ q 1 , q 2 ,…q s ] the initial probability distribution for the Markov chain. In most applications, the transition probabilities are displayed as an s x s transition probability matrix P . The transition probability matrix P may be written as
  • 9. For each I We also know that each entry in the P matrix must be nonnegative. Hence, all entries in the transition probability matrix are nonnegative, and the entries in each row must sum to 1.
  • 10. The Gambler’s Ruin Problem At time 0, I have $2. At times 1, 2, …, I play a game in which I bet $1, with probabilities p, I win the game, and with probability 1 – p, I lose the game. My goal is to increase my capital to $4, and as soon as I do, the game is over. The game is also over if my capital is reduced to 0. Let X t represent my capital position after the time t game (if any) is played X 0 , X 1 , X 2 , …. May be viewed as a discrete-time stochastic process
  • 11. The Gambler’s Ruin Problem $0 $1 $2 $3 $4
  • 12. 5.3 n -Step Transition Probabilities A question of interest when studying a Markov chain is: If a Markov chain is in a state i at time m , what is the probability that n periods later the Markov chain will be in state j ? This probability will be independent of m , so we may write P ( X m+n = j | X m = i ) = P ( X n = j | X 0 = i ) = P ij ( n ) where P ij ( n ) is called the n -step probability of a transition from state i to state j . For n > 1, P ij ( n ) = ij th element of P n P ij (2) is the (i, j)th element of matrix P 2 = P 1 P 1 P ij (n) is the (i, j)th element of matrix P n = P 1 P n-1
  • 13. The Cola Example Suppose the entire cola industry produces only two colas. Given that a person last purchased cola 1, there is a 90% chance that their next purchase will be cola 1. Given that a person last purchased cola 2, there is an 80% chance that their next purchase will be cola 2. If a person is currently a cola 2 purchaser, what is the probability that they will purchase cola 1 two purchases from now? If a person is currently a cola 1 a purchaser, what is the probability that they will purchase cola 1 three purchases from now?
  • 14. The Cola Example We view each person’s purchases as a Markov chain with the state at any given time being the type of cola the person last purchased. Hence, each person’s cola purchases may be represented by a two-state Markov chain, where State 1 = person has last purchased cola 1 State 2 = person has last purchased cola 2 If we define X n to be the type of cola purchased by a person on her n th future cola purchase, then X 0 , X 1 , … may be described as the Markov chain with the following transition matrix:
  • 15. The Cola Example We can now answer questions 1 and 2. We seek P ( X 2 = 1| X 0 = 2) = P 21 (2) = element 21 of P 2 :
  • 16. The Cola Example Hence, P 21 (2) =.34. This means that the probability is .34 that two purchases in the future a cola 2 drinker will purchase cola 1. We seek P 11 (3) = element 11 of P 3 : Therefore, P 11 (3) = .781
  • 17. Many times we do not know the state of the Markov chain at time 0. Then we can determine the probability that the system is in state i at time n by using the reasoning. Probability of being in state j at time n where q =[q 1 , q 2 , … q 3 ]. Hence, q n = q o p n = q n-1 p Example, q 0 = (.4,.6) q 1 = (.4,.6) q 1 = (.48,.52)
  • 18. To illustrate the behavior of the n -step transition probabilities for large values of n , we have computed several of the n -step transition probabilities for the Cola example. This means that for large n, no matter what the initial state, there is a .67 chance that a person will be a cola 1 purchaser.
  • 19. 5.4 Classification of States in a Markov Chain To understand the n -step transition in more detail, we need to study how mathematicians classify the states of a Markov chain. The following transition matrix illustrates most of the following definitions. A graphical representation is shown in the book (State-Transition diagram)
  • 20. Definition: Given two states of i and j , a path from i to j is a sequence of transitions that begins in i and ends in j , such that each transition in the sequence has a positive probability of occurring. Definition: A state j is reachable from state i if there is a path leading from i to j . Definition: Two states i and j are said to communicate if j is reachable from i , and i is reachable from j . Definition: A set of states S in a Markov chain is a closed set if no state outside of S is reachable from any state in S .
  • 21. Definition: A state i is an absorbing state if p ij =1. Definition: A state i is a transient state if there exists a state j that is reachable from i , but the state i is not reachable from state j . Definition: If a state is not transient, it is called a recurrent state. Definition: A state i is periodic with period k > 1 if k is the smallest number such that all paths leading from state i back to state i have a length that is a multiple of k. If a recurrent state is not periodic, it is referred to as aperiodic . If all states in a chain are recurrent, aperiodic, and communicate with each other, the chain is said to be ergodic . The importance of these concepts will become clear after the next two sections.
  • 22. 5.5 Steady-State Probabilities and Mean First Passage Times Steady-state probabilities are used to describe the long-run behavior of a Markov chain. Theorem 1: Let P be the transition matrix for an s -state ergodic chain. Then there exists a vector π = [ π 1 π 2 … π s ] such that
  • 23. Theorem 1 tells us that for any initial state i , The vector π = [ π 1 π 2 … π s ] is often called the steady-state distribution , or equilibrium distribution , for the Markov chain. Hence, they are independent of the initial probability distribution defined over the states
  • 24. Transient Analysis & Intuitive Interpretation The behavior of a Markov chain before the steady state is reached is often call transient (or short-run) behavior . An interpretation can be given to the steady-state probability equations. This equation may be viewed as saying that in the steady-state, the “flow” of probability into each state must equal the flow of probability out of each state.
  • 25. Steady-State Probabilities The vector  = [  1 ,  2 , …. ,  s ] is often known as the steady-state distribution for the Markov chain For large n and all i, P ij (n+1)  P ij (n)   j In matrix form  =  P For any n and any i, P i1 (n) + P i2 (n) + … + P is (n) = 1 As n  , we have  1 +  2 + …. +  s = 1
  • 26. An Intuitive Interpretation of Steady-State Probabilities Consider Subtracting  j p jj from both sides of the above equation, we have Probability that a particular transition enters state j = probability that a particular transition leaves state j
  • 27. Use of Steady-State Probabilities in Decision Making In the Cola Example, suppose that each customer makes one purchase of cola during any week. Suppose there are 100 million cola customers. One selling unit of cola costs the company $1 to produce and is sold for $2. For $500 million/year, an advertising firm guarantees to decrease from 10% to 5% the fraction of cola 1 customers who switch after a purchase. Should the company that makes cola 1 hire the firm?
  • 28. At present, a fraction π 1 = ⅔ of all purchases are cola 1 purchases, since: π 1 = .90 π 1 +.20 π 2 π 2 = .10 π 1 +.80 π 2 and using the following equation by π 1 + π 2 = 1 Each purchase of cola 1 earns the company a $1 profit. We can calculate the annual profit as $3,466,666,667 [2/3(100 million)(52 weeks)$1]. The advertising firm is offering to change the P matrix to
  • 29. For P 1, the steady-state equations become π 1 = .95 π 1 +.20 π 2 π 2 = .05 π 1 +.80 π 2 Replacing the second equation by π 1 + π 2 = 1 and solving, we obtain π 1 =.8 and π 2 = .2. Now the cola 1 company’s annual profit will be $3,660,000,000 [.8(100 million)(52 weeks)$1-($500 million)]. Hence, the cola 1 company should hire the ad agency.
  • 30. Inventory Example A camera store stocks a particular model camera that can be ordered weekly. Let D 1 , D 2 , … represent the demand for this camera (the number of units that would be sold if the inventory is not depleted) during the first week, second week, …, respectively. It is assumed that the D i ’s are independent and identically distributed random variables having a Poisson distribution with a mean of 1. Let X 0 represent the number of cameras on hand at the outset, X 1 the number of cameras on hand at the end of week 1, X 2 the number of cameras on hand at the end of week 2, and so on. Assume that X 0 = 3. On Saturday night the store places an order that is delivered in time for the next opening of the store on Monday. The store using the following order policy: If there are no cameras in stock, 3 cameras are ordered. Otherwise, no order is placed. Sales are lost when demand exceeds the inventory on hand
  • 31. Inventory Example X t is the number of Cameras in stock at the end of week t (as defined earlier), where X t represents the state of the system at time t Given that X t = i, X t+1 depends only on D t+1 and X t (Markovian property) D t has a Poisson distribution with mean equal to one. This means that P(D t+1 = n) = e -1 1 n /n! for n = 0, 1, … P(D t = 0 ) = e -1 = 0.368 P(D t = 1 ) = e -1 = 0.368 P(D t = 2 ) = (1/2)e -1 = 0.184 P(D t  3 ) = 1 – P(D t  2) = 1 – (.368 + .368 + .184) = 0.08 X t+1 = max(3-D t+1 , 0) if X t = 0 and X t+1 = max(X t – D t+1 , 0) if X t  1, for t = 0, 1, 2, ….
  • 32. Inventory Example: (One-Step) Transition Matrix P 03 = P(D t+1 = 0) = 0.368 P 02 = P(D t+1 = 1) = 0.368 P 01 = P(D t+1 = 2) = 0.184 P 00 = P(D t+1  3) = 0.080
  • 33. Inventory Example: Transition Diagram 0 1 2 3
  • 34. Inventory Example: (One-Step) Transition Matrix
  • 36. Transition Matrix: Four-Step P (4) = P (2) P (2)
  • 37. Transition Matrix: Eight-Step P (8) = P (4) P (4)
  • 38. Steady-State Probabilities The steady-state probabilities uniquely satisfy the following steady-state equations  0 =  0 p 00 +  1 p 10 +  2 p 20 +  3 p 30  1 =  0 p 01 +  1 p 11 +  2 p 21 +  3 p 31  2 =  0 p 02 +  1 p 12 +  2 p 22 +  3 p 32  3 =  0 p 03 +  1 p 13 +  2 p 23 +  3 p 33 1 =  0 +  1 +  2 +  3
  • 39. Steady-State Probabilities: Inventory Example  0 = .080  0 + .632  1 + .264  2 + .080  3  1 = .184  0 + .368  1 + .368  2 + .184  3  2 = .368  0 + .368  2 + .368  3  3 = .368  0 + .368  3 1 =  0 +  1 +  2 +  3  0 = .286,  1 = .285,  2 = .263,  3 = .166 The numbers in each row of matrix P (8) match the corresponding steady-state probability
  • 40. Mean First Passage Times For an ergodic chain, let m ij = expected number of transitions before we first reach state j , given that we are currently in state i; m ij is called the mean first passage time from state i to state j . In the example, we assume we are currently in state i . Then with probability p ij , it will take one transition to go from state i to state j . For k ≠ j , we next go with probability p ik to state k . In this case, it will take an average of 1 + m kj transitions to go from i and j .
  • 41. This reasoning implies By solving the linear equations of the equation above, we find all the mean first passage times. It can be shown that
  • 42. For the cola example, π 1 =2/3 and π 2 = 1/3 Hence, m 11 = 1.5 and m 22 = 3 m 12 = 1 + p 11 m 12 = 1 + .9m 12 m 21 = 1 + p 22 m 21 = 1 + .8m 21 Solving these two equations yields, m 12 = 10 and m 21 = 5
  • 43. Solving for Steady-State Probabilities and Mean First Passage Times on the Computer Since we solve steady-state probabilities and mean first passage times by solving a system of linear equations, we may use LINDO to determine them. Simply type in an objective function of 0, and type the equations you need to solve as your constraints.
  • 44. 5.6 Absorbing Chains Many interesting applications of Markov chains involve chains in which some of the states are absorbing and the rest are transient states. This type of chain is called an absorbing chain . To see why we are interested in absorbing chains we consider the following accounts receivable example.
  • 45. Accounts Receivable Example The accounts receivable situation of a firm is often modeled as an absorbing Markov chain. Suppose a firm assumes that an account is uncollected if the account is more than three months overdue. Then at the beginning of each month, each account may be classified into one of the following states: State 1 New account State 2 Payment on account is one month overdue State 3 Payment on account is two months overdue State 4 Payment on account is three months overdue State 5 Account has been paid State 6 Account is written off as bad debt
  • 46. Suppose that past data indicate that the following Markov chain describes how the status of an account changes from one month to the next month: New 1 month 2 months 3 months Paid Bad Debt New 1 month 2 months 3 months Paid Bad Debt
  • 47. To simplify our example, we assume that after three months, a debt is either collected or written off as a bad debt. Once a debt is paid up or written off as a bad debt, the account if closed, and no further transitions occur. Hence, Paid or Bad Debt are absorbing states. Since every account will eventually be paid or written off as a bad debt, New, 1 month, 2 months, and 3 months are transient states.
  • 48. A typical new account will be absorbed as either a collected debt or a bad debt. What is the probability that a new account will eventually be collected? To answer this questions we must write a transition matrix. We assume s – m transient states and m absorbing states. The transition matrix is written in the form of m columns s-m rows m rows s-m columns P =
  • 49. The transition matrix for this example is Then s =6, m =2, and Q and R are as shown. New 1 month 2 months 3 months Paid Bad Debt New 1 month 2 months 3 months Paid Bad Debt Q R
  • 50. What is the probability that a new account will eventually be collected? (.964) What is the probability that a one-month overdue account will eventually become a bad debt? (.06) If the firm’s sales average $100,000 per month, how much money per year will go uncollected? From answer 1, only 3.6% of all debts are uncollected. Since yearly accounts payable are $1,200,000 on the average, (0.036)(1,200,000) = $43,200 per year will be uncollected.