Fiveable

📊Actuarial Mathematics Unit 2 Review

QR code for Actuarial Mathematics practice questions

2.1 Markov chains and transition probabilities

2.1 Markov chains and transition probabilities

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📊Actuarial Mathematics
Unit & Topic Study Guides

Markov chains are a core modeling tool in actuarial mathematics for systems that evolve over time through discrete states. By capturing how a system moves between states using transition probabilities, they let actuaries forecast outcomes for insurance portfolios, claims processes, and financial risk models.

This topic covers the formal structure of Markov chains, how to compute and interpret transition probabilities, state classification, long-term behavior, first passage and absorption analysis, estimation methods, and Markov decision processes.

Definition of Markov chains

A Markov chain is a stochastic process that transitions between a set of states over time, where the probability of each transition depends only on the current state. This "memoryless" behavior is what makes Markov chains both tractable and widely applicable.

In actuarial work, Markov chains model phenomena like policyholder behavior, claim frequency patterns, and credit rating migrations.

States in Markov chains

A state represents a possible condition the system can occupy at a given time. The collection of all possible states is the state space, which can be finite or countably infinite.

Common actuarial examples:

  • Policyholder status: active, lapsed, or surrendered
  • Claim status: no claim, minor claim, or major claim
  • Credit ratings: AAA, AA, A, BBB, and so on down to default

Discrete vs continuous time

  • Discrete-time Markov chains (DTMCs) have transitions occurring at fixed intervals (e.g., monthly or annually). These are the primary focus of this topic.
  • Continuous-time Markov chains (CTMCs) allow transitions at any moment. The time spent in each state before transitioning follows an exponential distribution, which itself has the memoryless property.

Memoryless property

The Markov property (memoryless property) is the defining feature of a Markov chain. It says that the future state depends only on the present state, not on how the system arrived there.

Formally, for states ii, jj, kk, ..., ll and time steps nn and mm:

P(Xn+m=jXn=i,Xn1=k,,X0=l)=P(Xn+m=jXn=i)P(X_{n+m} = j \mid X_n = i, X_{n-1} = k, \ldots, X_0 = l) = P(X_{n+m} = j \mid X_n = i)

This is what makes the math manageable: you only need to know where the system is now to compute where it goes next.

Transition probabilities

Transition probabilities quantify how likely the system is to move from one state to another. They're the building blocks for every calculation you'll do with Markov chains.

One-step transition probabilities

The one-step transition probability pijp_{ij} is the probability of moving from state ii to state jj in a single time step. For a DTMC with NN states, these probabilities fill an N×NN \times N transition probability matrix P\mathbf{P}.

Each entry pijp_{ij} satisfies pij0p_{ij} \geq 0, and each row sums to 1 (since the system must go somewhere).

Multi-step transition probabilities

The multi-step transition probability pij(n)p_{ij}^{(n)} is the probability of going from state ii to state jj in exactly nn steps. You compute it by raising the transition matrix to the nn-th power:

P(n)=Pn\mathbf{P}^{(n)} = \mathbf{P}^n

The Chapman-Kolmogorov equations give a recursive way to break this computation apart. For any intermediate point mm:

pij(n+m)=k=1Npik(n)pkj(m)p_{ij}^{(n+m)} = \sum_{k=1}^{N} p_{ik}^{(n)} \, p_{kj}^{(m)}

This says: to get from ii to jj in n+mn+m steps, sum over all possible intermediate states kk that the chain could pass through at step nn.

Transition probability matrix

The transition probability matrix P\mathbf{P} is a square matrix where row ii, column jj contains pijp_{ij}. Two properties always hold:

  • Every entry is non-negative: pij0p_{ij} \geq 0
  • Every row sums to 1: j=1Npij=1\sum_{j=1}^{N} p_{ij} = 1

A matrix satisfying these conditions is called a stochastic matrix. When you see a transition matrix on an exam, check these two properties first to verify it's valid.

Classification of states

How states behave in the long run depends on their structural relationships within the chain. Classifying states tells you whether the chain settles down, cycles, or gets trapped.

Communicating states

States ii and jj communicate (written iji \leftrightarrow j) if you can reach jj from ii and reach ii from jj, each in a finite number of steps.

Communication is an equivalence relation, so it partitions the state space into communicating classes. If the entire chain forms a single communicating class (every state can reach every other state), the chain is called irreducible.

Absorbing states

An absorbing state is one you can never leave: pii=1p_{ii} = 1. Once the chain enters an absorbing state, it stays there forever.

A chain with at least one absorbing state and the property that every transient state can eventually reach some absorbing state is called an absorbing Markov chain. Actuarial examples include policy termination or death of a policyholder, where the state is permanent.

Transient vs recurrent states

  • A state is transient if there's a positive probability of never returning to it after leaving.
  • A state is recurrent if the chain is guaranteed (probability 1) to return to it eventually.

Recurrent states break down further:

  • Positive recurrent: The expected return time is finite. These are the states that matter for long-run analysis.
  • Null recurrent: The chain returns with probability 1, but the expected return time is infinite. This only occurs in chains with infinitely many states.

In a finite, irreducible chain, all states are positive recurrent.

Long-term behavior

Understanding where a Markov chain "settles" over time is critical for actuarial forecasting. The key question: does the chain converge to a stable distribution?

Stationary distribution

A stationary distribution π\boldsymbol{\pi} is a probability vector that satisfies:

π=πP\boldsymbol{\pi} = \boldsymbol{\pi} \mathbf{P}

If the chain starts in the distribution π\boldsymbol{\pi}, it stays in that distribution after every transition. Think of it as an equilibrium: the inflows and outflows for each state are perfectly balanced.

Existence and uniqueness depend on the chain's properties. An irreducible, positive recurrent chain always has a unique stationary distribution.

Limiting distribution

The limiting distribution π\boldsymbol{\pi}^* describes the long-run proportion of time spent in each state, regardless of where the chain started. For an irreducible and aperiodic chain, the limiting distribution exists, is unique, and equals the stationary distribution.

To find it, solve:

π=πP\boldsymbol{\pi}^* = \boldsymbol{\pi}^* \mathbf{P}

i=1Nπi=1\sum_{i=1}^{N} \pi_i^* = 1

The first equation gives you NN equations (one per state, though one is redundant), and the second ensures the result is a valid probability distribution.

Aperiodicity matters. If a chain is periodic (e.g., it cycles between states with a fixed period), the limiting distribution doesn't exist even though a stationary distribution does. The chain keeps cycling instead of settling down.

Convergence of Markov chains

For an irreducible, aperiodic chain, the state probabilities converge to the limiting distribution over time, no matter the starting state. The rate of convergence is governed by the second-largest eigenvalue (in absolute value) of P\mathbf{P}. The closer this eigenvalue is to 0, the faster the chain converges.

This matters in practice: slow convergence means you need more time steps before the long-run distribution is a reliable approximation.

First passage times

First passage times measure how long it takes the chain to reach a target state for the first time. In actuarial contexts, this could represent the time until a first claim, or the number of periods until a policyholder lapses.

Definition of first passage time

The first passage time from state ii to state jj, denoted TijT_{ij}, is:

Tij=min{n1:Xn=jX0=i}T_{ij} = \min\{n \geq 1 : X_n = j \mid X_0 = i\}

This is a random variable. When i=ji = j, it's called the first return time to state ii.

States in Markov chains, Markov chain - Dave Tang's blog

Expected first passage times

The expected first passage time mijm_{ij} is the average number of steps to first reach state jj from state ii. You find it by solving:

mij=1+kjpikmkjm_{ij} = 1 + \sum_{k \neq j} p_{ik} \, m_{kj}

The logic: from state ii, you take one step (the "1"), and if you don't land on jj, you end up in some state kk and still need mkjm_{kj} more steps on average. This gives a system of linear equations you solve simultaneously for all starting states.

Variance of first passage times

The variance vijv_{ij} measures how spread out the first passage time is around its mean. It's computed from:

vij=kjpik(vkj+mkj2)mij2v_{ij} = \sum_{k \neq j} p_{ik} \left( v_{kj} + m_{kj}^2 \right) - m_{ij}^2

You need the expected first passage times mkjm_{kj} before you can solve for the variances. Together, the mean and variance give a much more complete picture of the first passage time distribution.

Absorption probabilities

For absorbing Markov chains, a central question is: starting from a transient state, what's the probability of ending up in each absorbing state?

Calculation of absorption probabilities

The absorption probability aija_{ij} is the probability that the chain, starting in transient state ii, eventually gets absorbed into absorbing state jj. Solve:

aij=pij+kTpikakja_{ij} = p_{ij} + \sum_{k \in T} p_{ik} \, a_{kj}

where TT is the set of transient states. This is another system of linear equations: one equation per transient state, for each absorbing state.

Fundamental matrix

The fundamental matrix N\mathbf{N} encodes the expected number of times the chain visits each transient state before being absorbed:

N=(IQ)1\mathbf{N} = (\mathbf{I} - \mathbf{Q})^{-1}

Here, Q\mathbf{Q} is the submatrix of P\mathbf{P} containing only transitions among transient states, and I\mathbf{I} is the identity matrix.

The entry nijn_{ij} gives the expected number of visits to transient state jj before absorption, starting from transient state ii. This matrix is the workhorse for absorption calculations.

Time to absorption

The expected absorption time from transient state ii is simply the sum of the ii-th row of the fundamental matrix:

ti=jTnijt_i = \sum_{j \in T} n_{ij}

This makes intuitive sense: the total expected time before absorption equals the sum of expected visits to every transient state. The variance of absorption time can also be derived from N\mathbf{N}.

Applications of Markov chains

Markov chains appear throughout actuarial practice. Three classical applications illustrate different aspects of the theory.

Queueing theory

Queueing models describe systems with arrivals and service (e.g., claims arriving at a processing center). The states represent the number of items in the queue, and transition probabilities depend on arrival and service rates.

By finding the steady-state distribution, actuaries can determine average queue lengths, waiting times, and the probability of exceeding capacity. This informs staffing and resource allocation decisions.

Gambler's ruin problem

A gambler starts with capital kk and plays repeated games, winning or losing one unit each round with fixed probabilities pp and q=1pq = 1 - p. The states are the gambler's current capital (0 through some upper bound NN), with 0 and NN as absorbing states.

This is a classic absorbing chain problem. The absorption probabilities give the probability of ruin (reaching 0) versus reaching the target NN, and the expected absorption time gives the expected duration of play. The same framework applies to surplus models in ruin theory.

Birth-death processes

Birth-death processes are CTMCs where transitions only occur between neighboring states: the population can increase by one (birth) or decrease by one (death) at any time.

In actuarial work, these model portfolio sizes (new policies vs. lapses), claim counts over time, or disease spread in health insurance populations. The balance between birth and death rates determines whether the system has a stable steady-state distribution or grows/shrinks without bound.

Estimation of transition probabilities

In practice, you rarely know the true transition probabilities. You estimate them from observed data, such as historical records of policyholder transitions or credit rating changes.

Maximum likelihood estimation

Maximum likelihood estimation (MLE) is the standard approach. Given observed transitions, the MLE for each transition probability is:

p^ij=nijni\hat{p}_{ij} = \frac{n_{ij}}{n_i}

where nijn_{ij} is the number of observed transitions from state ii to state jj, and nin_i is the total number of transitions out of state ii. This is simply the observed proportion, and it automatically satisfies the row-sum constraint.

Bayesian estimation

Bayesian estimation incorporates prior beliefs about the transition probabilities before seeing data. A common choice is a Dirichlet prior for each row of the transition matrix, since it's the conjugate prior for multinomial data.

The posterior distribution combines the prior with the observed transition counts via Bayes' theorem. This approach is especially useful when data is sparse, as the prior prevents extreme estimates, and it naturally quantifies parameter uncertainty through the full posterior distribution.

Confidence intervals for probabilities

Confidence intervals quantify uncertainty in the estimated transition probabilities. Two main approaches:

  • Wald intervals: Use the asymptotic normality of the MLE. The standard error for p^ij\hat{p}_{ij} is approximately p^ij(1p^ij)/ni\sqrt{\hat{p}_{ij}(1 - \hat{p}_{ij}) / n_i}, and a 95% interval is p^ij±1.96×SE\hat{p}_{ij} \pm 1.96 \times SE.
  • Bootstrap methods: Resample from the observed data, re-estimate the transition matrix each time, and use the empirical distribution of estimates to construct intervals.

Bootstrap methods tend to perform better when sample sizes are small or when p^ij\hat{p}_{ij} is near 0 or 1, where the normal approximation breaks down.

Markov decision processes

Markov decision processes (MDPs) extend Markov chains by adding a decision-maker who can choose actions that influence both the transitions and the rewards received. They're the foundation for optimal sequential decision-making under uncertainty.

States, actions, and rewards

An MDP is defined by four components:

  • States: The possible conditions of the system (same as in a Markov chain)
  • Actions: At each state, the decision-maker selects from a set of available actions
  • Transition probabilities: Now depend on both the current state and the chosen action: pij(a)p_{ij}(a)
  • Rewards: Each state-action pair yields an immediate reward r(i,a)r(i, a), representing costs or benefits

The objective is to find a strategy (policy) that maximizes the expected total discounted reward over time.

Optimal policies

A policy maps each state to an action. The optimal policy maximizes the expected total reward, accounting for both immediate and future consequences.

The Bellman equation expresses this recursively. For the optimal value function V(i)V^*(i):

V(i)=maxa[r(i,a)+γjpij(a)V(j)]V^*(i) = \max_{a} \left[ r(i, a) + \gamma \sum_{j} p_{ij}(a) \, V^*(j) \right]

where γ\gamma is the discount factor (0γ<10 \leq \gamma < 1). The optimal action at each state is whichever action achieves this maximum.

Value iteration vs policy iteration

Two standard algorithms solve for the optimal policy:

Value iteration:

  1. Start with an initial guess for V(i)V(i) (e.g., all zeros)
  2. Update each state's value using the Bellman equation
  3. Repeat until the values converge (changes fall below a threshold)

Policy iteration:

  1. Start with an arbitrary policy
  2. Policy evaluation: Compute the value function for the current policy by solving a system of linear equations
  3. Policy improvement: Update the policy by choosing the action that maximizes value at each state
  4. Repeat steps 2-3 until the policy stops changing

Policy iteration typically converges in fewer iterations but each iteration is more expensive (it requires solving a linear system). Value iteration does simpler updates but may need more iterations. For large state spaces with sparse transitions, value iteration is often preferred.