All Study Guides Stochastic Processes Unit 5
🔀 Stochastic Processes Unit 5 – Markov chainsMarkov chains are mathematical models that describe systems transitioning between states over time. They're used to analyze everything from web page rankings to queueing systems, relying on the Markov property that future states depend only on the current state.
Key concepts include state spaces, transition probabilities, and stationary distributions. Different types of Markov chains exist, such as finite-state, absorbing, and ergodic chains. Understanding these helps in applying Markov chains to real-world problems and developing computational methods for analysis.
Key Concepts and Definitions
Markov chains model systems transitioning between discrete states over time
Markov property assumes future states depend only on the current state, not past states
State space S S S represents the set of all possible states in the system
Time index t t t can be discrete (e.g., t = 0 , 1 , 2 , … t=0,1,2,\ldots t = 0 , 1 , 2 , … ) or continuous (e.g., t ≥ 0 t \geq 0 t ≥ 0 )
Discrete-time Markov chains (DTMCs) have discrete time steps
Continuous-time Markov chains (CTMCs) have continuous time
Initial distribution π 0 \pi_0 π 0 specifies the probability of starting in each state
Transition probabilities p i j p_{ij} p ij define the likelihood of moving from state i i i to state j j j
Markov chains can be homogeneous (transition probabilities remain constant over time) or inhomogeneous (transition probabilities change with time)
Types of Markov Chains
Finite-state Markov chains have a finite number of states in the state space
Infinite-state Markov chains have an infinite number of states (e.g., birth-death processes)
Absorbing Markov chains contain at least one absorbing state that, once entered, cannot be left
Ergodic Markov chains are irreducible and aperiodic, ensuring a unique stationary distribution exists
Irreducible Markov chains allow reaching any state from any other state in a finite number of steps
Aperiodic Markov chains do not have periodic behavior in state transitions
Time-homogeneous Markov chains have transition probabilities that remain constant over time
Time-inhomogeneous Markov chains have transition probabilities that vary with time
Higher-order Markov chains consider multiple previous states to determine the next state
Transition Probabilities and Matrices
Transition probabilities p i j p_{ij} p ij represent the probability of moving from state i i i to state j j j in one step
p i j = P ( X t + 1 = j ∣ X t = i ) p_{ij} = P(X_{t+1} = j | X_t = i) p ij = P ( X t + 1 = j ∣ X t = i ) , where X t X_t X t is the state at time t t t
Transition probabilities satisfy ∑ j p i j = 1 \sum_{j} p_{ij} = 1 ∑ j p ij = 1 for each state i i i
Transition probability matrix P P P contains all transition probabilities p i j p_{ij} p ij
Rows of P P P sum to 1, as they represent probability distributions
n n n -step transition probabilities p i j ( n ) p_{ij}^{(n)} p ij ( n ) give the probability of moving from state i i i to state j j j in exactly n n n steps
Can be computed using matrix multiplication: P ( n ) = P n P^{(n)} = P^n P ( n ) = P n
Limiting probabilities π j = lim n → ∞ p i j ( n ) \pi_j = \lim_{n \to \infty} p_{ij}^{(n)} π j = lim n → ∞ p ij ( n ) describe the long-term behavior of the chain
Transition rate matrix Q Q Q is used for continuous-time Markov chains
Off-diagonal entries q i j q_{ij} q ij represent the rate of transitioning from state i i i to state j j j
Diagonal entries q i i q_{ii} q ii are set such that each row sums to 0
State Classification and Properties
Communicating states can be reached from each other in a finite number of steps
Recurrent states are visited infinitely often with probability 1
Positive recurrent states have a finite expected return time
Null recurrent states have an infinite expected return time
Transient states are visited only finitely many times with probability 1
Absorbing states are recurrent states that, once entered, cannot be left
Periodicity of a state is the greatest common divisor of the return times to that state
Aperiodic states have a period of 1
Ergodic Markov chains are irreducible and aperiodic
All states communicate with each other and are aperiodic
Detailed balance equations relate stationary probabilities and transition rates in reversible Markov chains
Stationary Distributions
Stationary distribution π \pi π is a probability vector that remains unchanged under the transition matrix
Satisfies π P = π \pi P = \pi π P = π for discrete-time Markov chains
Satisfies π Q = 0 \pi Q = 0 π Q = 0 for continuous-time Markov chains
Stationary distributions represent the long-term behavior of the Markov chain
Ergodic Markov chains have a unique stationary distribution
Can be found by solving the linear system π P = π \pi P = \pi π P = π or π Q = 0 \pi Q = 0 π Q = 0
Time-reversible Markov chains have the same stationary distribution when time is reversed
Satisfy the detailed balance equations π i p i j = π j p j i \pi_i p_{ij} = \pi_j p_{ji} π i p ij = π j p ji or π i q i j = π j q j i \pi_i q_{ij} = \pi_j q_{ji} π i q ij = π j q ji
Mixing time measures how quickly the chain converges to its stationary distribution
Convergence rate depends on the spectral gap of the transition matrix or generator
Applications and Examples
PageRank algorithm uses a Markov chain to rank web pages in search engine results
States represent web pages, and transitions represent hyperlinks between pages
Queueing systems can be modeled using Markov chains (e.g., M/M/1 queue)
States represent the number of customers in the system
Transitions occur when customers arrive or depart
Hidden Markov models (HMMs) are used in speech recognition and bioinformatics
Observable states emit symbols according to an emission probability distribution
Hidden states form a Markov chain governing the transitions between observable states
Markov decision processes (MDPs) combine Markov chains with decision-making for reinforcement learning
Actions taken in each state influence the transition probabilities and rewards
Markov chain Monte Carlo (MCMC) methods sample from complex probability distributions
Construct a Markov chain whose stationary distribution is the target distribution
Examples include Metropolis-Hastings algorithm and Gibbs sampling
Computational Methods
Matrix multiplication can be used to compute n n n -step transition probabilities: P ( n ) = P n P^{(n)} = P^n P ( n ) = P n
Solving linear systems π P = π \pi P = \pi π P = π or π Q = 0 \pi Q = 0 π Q = 0 yields stationary distributions
Gaussian elimination, iterative methods (e.g., Jacobi, Gauss-Seidel), or specialized algorithms can be used
Eigenvalue decomposition of the transition matrix provides insights into long-term behavior
Largest eigenvalue is always 1, corresponding to the stationary distribution
Spectral gap between the largest and second-largest eigenvalues determines the convergence rate
Simulation techniques generate sample paths of the Markov chain
Useful for estimating probabilities, expected values, and other quantities of interest
Numerical methods for continuous-time Markov chains include uniformization and Runge-Kutta methods
Specialized software packages (e.g., PRISM, PyMC, R packages) facilitate Markov chain analysis
Advanced Topics and Extensions
Markov renewal processes generalize Markov chains by allowing non-exponential sojourn times in each state
Semi-Markov processes have state transitions that depend on the sojourn time in the current state
Markov-modulated processes use a Markov chain to govern the parameters of another stochastic process
Examples include Markov-modulated Poisson processes and Markov-modulated fluid flows
Markov random fields extend Markov chains to higher dimensions (e.g., image processing, spatial statistics)
States are associated with nodes in a graph, and transitions occur between neighboring nodes
Partially observable Markov decision processes (POMDPs) introduce uncertainty in the state observations
Decision-making must account for the belief state, a probability distribution over the possible states
Reinforcement learning algorithms (e.g., Q-learning, SARSA) use Markov decision processes to learn optimal policies
Convergence analysis and mixing times are essential for understanding the long-term behavior of Markov chains
Spectral methods, coupling arguments, and Lyapunov functions are used to establish convergence rates