Fiveable

๐Ÿ”€Stochastic Processes Unit 6 Review

QR code for Stochastic Processes practice questions

6.2 Infinitesimal generator matrix

6.2 Infinitesimal generator matrix

Written by the Fiveable Content Team โ€ข Last updated August 2025
Written by the Fiveable Content Team โ€ข Last updated August 2025
๐Ÿ”€Stochastic Processes
Unit & Topic Study Guides

Definition of the Infinitesimal Generator Matrix

The infinitesimal generator matrix, denoted QQ, encodes all the information about how a continuous-time Markov chain (CTMC) evolves. Where a discrete-time chain uses a transition probability matrix to describe jumps at fixed time steps, the generator matrix captures instantaneous rates of transition between states.

Each entry qijq_{ij} (with iโ‰ ji \neq j) represents the rate at which the process transitions from state ii to state jj over an infinitesimal time interval. More precisely, for small ฮ”t\Delta t:

P(X(t+ฮ”t)=jโˆฃX(t)=i)โ‰ˆqijโ€‰ฮ”t,iโ‰ jP(X(t + \Delta t) = j \mid X(t) = i) \approx q_{ij} \, \Delta t, \quad i \neq j

The diagonal entries qiiq_{ii} are defined so that each row sums to zero, which means qii=โˆ’โˆ‘jโ‰ iqijq_{ii} = -\sum_{j \neq i} q_{ij}. This makes โˆฃqiiโˆฃ|q_{ii}| the total rate at which the process leaves state ii.

Properties of the Infinitesimal Generator Matrix

Non-negative off-diagonal elements

Off-diagonal entries qijq_{ij} (for iโ‰ ji \neq j) are always non-negative. They represent transition rates, and a negative rate has no physical meaning. If qij=0q_{ij} = 0, direct transitions from ii to jj simply don't occur.

Negative diagonal elements

Each diagonal entry qiiq_{ii} is non-positive (negative or zero). Its absolute value โˆฃqiiโˆฃ|q_{ii}| gives the total departure rate from state ii. The holding time in state ii is exponentially distributed with parameter โˆฃqiiโˆฃ|q_{ii}|, so the expected time spent in state ii before jumping is 1/โˆฃqiiโˆฃ1/|q_{ii}|.

A diagonal entry of zero means the state is absorbing: once the process enters, it never leaves.

Row sums equal to zero

Every row of QQ sums to zero:

โˆ‘jqij=0forย allย i\sum_{j} q_{ij} = 0 \quad \text{for all } i

This is the continuous-time analogue of rows summing to one in a discrete-time transition matrix. It enforces conservation of probability: probability flowing out of state ii (captured by qiiq_{ii}) exactly balances the total probability flowing into other states (captured by the off-diagonal entries).

Relationship to the Transition Rate Matrix

Some texts define a separate transition rate matrix RR whose entries rijr_{ij} (for iโ‰ ji \neq j) are the same transition rates, but whose diagonal entries are zero (or defined differently). The generator QQ and RR share the same off-diagonal entries:

qij=rij,iโ‰ jq_{ij} = r_{ij}, \quad i \neq j

The diagonal of QQ is then constructed from RR:

qii=โˆ’โˆ‘jโ‰ irijq_{ii} = -\sum_{j \neq i} r_{ij}

So QQ is fully determined by the off-diagonal rates. If you're given RR, building QQ just means filling in the diagonal to make each row sum to zero.

Role in Continuous-Time Markov Chains

Evolution of state probabilities

Let p(t)\mathbf{p}(t) be the row vector of state probabilities at time tt. The generator matrix governs how p(t)\mathbf{p}(t) changes over time through the relationship:

ddtp(t)=p(t)โ€‰Q\frac{d}{dt}\mathbf{p}(t) = \mathbf{p}(t) \, Q

This single matrix equation encodes the entire dynamics of the CTMC. It also leads to the matrix exponential solution P(t)=eQtP(t) = e^{Qt}, where P(t)P(t) is the transition probability matrix at time tt.

Kolmogorov forward equations

The forward equations (also called master equations) describe how transition probabilities evolve forward in time:

ddtP(t)=P(t)โ€‰Q\frac{d}{dt} P(t) = P(t) \, Q

Given an initial distribution p(0)\mathbf{p}(0), you solve this system to find p(t)=p(0)โ€‰P(t)\mathbf{p}(t) = \mathbf{p}(0) \, P(t). These equations are the workhorse for computing state probabilities at any future time.

Kolmogorov backward equations

The backward equations take a different perspective, conditioning on the starting state:

ddtP(t)=Qโ€‰P(t)\frac{d}{dt} P(t) = Q \, P(t)

Notice QQ multiplies from the left here, not the right. The backward equations are particularly useful for computing hitting times and absorption probabilities, where you want to ask: "Starting from state ii, what's the probability of reaching state jj by time tt?"

Computation of the Infinitesimal Generator Matrix

From the transition rate matrix

  1. Write down all off-diagonal rates qij=rijq_{ij} = r_{ij} directly from the rate matrix RR.
  2. For each row ii, sum the off-diagonal entries: โˆ‘jโ‰ iqij\sum_{j \neq i} q_{ij}.
  3. Set the diagonal entry: qii=โˆ’โˆ‘jโ‰ iqijq_{ii} = -\sum_{j \neq i} q_{ij}.
  4. Verify every row sums to zero.

From a state transition diagram

  1. Label each state as a node. Each directed arrow from state ii to state jj carries a rate qijq_{ij}.
  2. Read off the arrow labels to fill in the off-diagonal entries of QQ.
  3. For each state ii, sum all outgoing arrow rates and negate to get qiiq_{ii}.

Example: Suppose you have three states with arrows: 1โ†’221 \xrightarrow{2} 2, 2โ†’312 \xrightarrow{3} 1, 2โ†’132 \xrightarrow{1} 3, and 3โ†’413 \xrightarrow{4} 1. The generator matrix is:

Q=(โˆ’2203โˆ’4140โˆ’4)Q = \begin{pmatrix} -2 & 2 & 0 \\ 3 & -4 & 1 \\ 4 & 0 & -4 \end{pmatrix}

Check: each row sums to zero.

Non-negative off-diagonal elements, Rotations And Infinitesimal Generators โ€“ Nathan Reedโ€™s coding blog

Eigenvalues and Eigenvectors

Interpretation of eigenvalues

The eigenvalues of QQ reveal how quickly the chain approaches equilibrium.

  • One eigenvalue is always ฮป1=0\lambda_1 = 0. This corresponds to the stationary distribution.
  • All other eigenvalues have non-positive real parts (for an irreducible, finite-state chain, they have strictly negative real parts).
  • The eigenvalue with the smallest nonzero magnitude (closest to zero) is called the spectral gap. It controls the rate of convergence to stationarity: a larger spectral gap means faster convergence.

Since P(t)=eQtP(t) = e^{Qt}, each eigenvalue ฮปk\lambda_k contributes a term proportional to eฮปkte^{\lambda_k t}. The zero eigenvalue gives the steady-state component, while the negative eigenvalues produce transient terms that decay exponentially.

Stationary distribution and eigenvectors

The stationary distribution ฯ€\boldsymbol{\pi} is the row vector satisfying:

ฯ€Q=0,โˆ‘iฯ€i=1\boldsymbol{\pi} Q = \mathbf{0}, \quad \sum_i \pi_i = 1

This makes ฯ€\boldsymbol{\pi} the left eigenvector of QQ corresponding to eigenvalue zero. Meanwhile, the right eigenvector for eigenvalue zero is 1\mathbf{1} (a column vector of ones), which is just a restatement of the row-sum-equals-zero property.

For an irreducible, positive-recurrent CTMC, this stationary distribution exists, is unique, and represents the long-run fraction of time spent in each state.

Applications of the Infinitesimal Generator Matrix

Queueing systems

In an M/M/1 queue with arrival rate ฮป\lambda and service rate ฮผ\mu, the states represent the number of customers in the system. The generator matrix has ฮป\lambda on the superdiagonal (arrivals) and ฮผ\mu on the subdiagonal (departures). Solving ฯ€Q=0\boldsymbol{\pi} Q = \mathbf{0} yields performance measures like average queue length and utilization ฯ=ฮป/ฮผ\rho = \lambda / \mu.

Birth-death processes

Birth-death processes are CTMCs where transitions only occur between neighboring states (iโ†’i+1i \to i+1 or iโ†’iโˆ’1i \to i-1). The generator matrix has a tridiagonal structure: birth rates ฮปi\lambda_i on the superdiagonal, death rates ฮผi\mu_i on the subdiagonal, and qii=โˆ’(ฮปi+ฮผi)q_{ii} = -(\lambda_i + \mu_i) on the diagonal. This structure makes the stationary distribution particularly tractable, with a product-form solution.

Reliability analysis

System components can be modeled as occupying operational or failed states, with failure rates and repair rates as the transition rates. The generator matrix captures all possible failure/repair transitions. From it, you can compute the mean time to failure (MTTF), system availability (long-run fraction of time the system is operational), and the probability of being in any particular degraded state.

Numerical Example

Consider a two-state CTMC (states 0 and 1) with transition rate ฮฑ\alpha from 0 to 1 and rate ฮฒ\beta from 1 to 0.

Step 1: Construct QQ:

Q=(โˆ’ฮฑฮฑฮฒโˆ’ฮฒ)Q = \begin{pmatrix} -\alpha & \alpha \\ \beta & -\beta \end{pmatrix}

Step 2: Find the stationary distribution by solving ฯ€Q=0\boldsymbol{\pi} Q = \mathbf{0} with ฯ€0+ฯ€1=1\pi_0 + \pi_1 = 1:

ฯ€0=ฮฒฮฑ+ฮฒ,ฯ€1=ฮฑฮฑ+ฮฒ\pi_0 = \frac{\beta}{\alpha + \beta}, \quad \pi_1 = \frac{\alpha}{\alpha + \beta}

Step 3: The eigenvalues are ฮป1=0\lambda_1 = 0 and ฮป2=โˆ’(ฮฑ+ฮฒ)\lambda_2 = -(\alpha + \beta). The transient behavior decays at rate ฮฑ+ฮฒ\alpha + \beta, so the chain reaches stationarity faster when both rates are large.

Step 4: The transition probability matrix is:

P(t)=eQt=1ฮฑ+ฮฒ(ฮฒ+ฮฑeโˆ’(ฮฑ+ฮฒ)tฮฑโˆ’ฮฑeโˆ’(ฮฑ+ฮฒ)tฮฒโˆ’ฮฒeโˆ’(ฮฑ+ฮฒ)tฮฑ+ฮฒeโˆ’(ฮฑ+ฮฒ)t)P(t) = e^{Qt} = \frac{1}{\alpha + \beta} \begin{pmatrix} \beta + \alpha e^{-(\alpha+\beta)t} & \alpha - \alpha e^{-(\alpha+\beta)t} \\ \beta - \beta e^{-(\alpha+\beta)t} & \alpha + \beta e^{-(\alpha+\beta)t} \end{pmatrix}

As tโ†’โˆžt \to \infty, the exponential terms vanish and each row converges to ฯ€\boldsymbol{\pi}.

Comparison to Discrete-Time Markov Chains

FeatureDTMCCTMC
Key matrixTransition probability matrix PPGenerator matrix QQ
Entry meaningpijp_{ij} = probability of jumping to jjqijq_{ij} = rate of transitioning to jj
Row sumsSum to 1Sum to 0
Entry constraintsAll entries in [0,1][0, 1]Off-diagonal โ‰ฅ0\geq 0, diagonal โ‰ค0\leq 0
Stationary conditionฯ€P=ฯ€\boldsymbol{\pi} P = \boldsymbol{\pi}ฯ€Q=0\boldsymbol{\pi} Q = \mathbf{0}
nn-step/time-tt evolutionPnP^n (matrix power)eQte^{Qt} (matrix exponential)
The connection between them: if you uniformize a CTMC (sample it at a rate at least as large as $$\max_iq_{ii}),yougetanequivalentDTMC.Conversely,), you get an equivalent DTMC. Conversely, Qcanbethoughtofasthe"derivative"ofthetransitionmatrixatcan be thought of as the "derivative" of the transition matrix att = 0:: Q = \lim_{\Delta t \to 0} \frac{P(\Delta t) - I}{\Delta t}$$.

Extensions and Generalizations

Time-inhomogeneous Markov chains

When transition rates change over time, the generator becomes a function Q(t)Q(t) with time-dependent entries qij(t)q_{ij}(t). The Kolmogorov equations still hold but with Q(t)Q(t) replacing QQ:

ddtP(s,t)=P(s,t)โ€‰Q(t)(forward)\frac{d}{dt} P(s, t) = P(s, t) \, Q(t) \quad \text{(forward)}

The solution is no longer a simple matrix exponential; instead it involves a time-ordered exponential (or product integral). This extension is useful for modeling systems with seasonal variation, aging components, or externally driven rate changes.

Absorbing states and absorption probabilities

An absorbing state aa has qaj=0q_{aj} = 0 for all jโ‰ aj \neq a (and therefore qaa=0q_{aa} = 0). To analyze absorption, partition the state space into transient states TT and absorbing states AA, and write QQ in block form:

Q=(QTQTA00)Q = \begin{pmatrix} Q_T & Q_{TA} \\ 0 & 0 \end{pmatrix}

Here QTQ_T is the submatrix of rates among transient states. Two key quantities follow:

  • Mean time to absorption from each transient state: m=โˆ’QTโˆ’11\mathbf{m} = -Q_T^{-1} \mathbf{1}
  • Absorption probabilities (probability of being absorbed into each absorbing state): B=โˆ’QTโˆ’1QTAB = -Q_T^{-1} Q_{TA}

These formulas are direct analogues of the discrete-time results using the fundamental matrix.