Fiveable

🔀Stochastic Processes Unit 4 Review

QR code for Stochastic Processes practice questions

4.3 Compound Poisson processes

4.3 Compound Poisson processes

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🔀Stochastic Processes
Unit & Topic Study Guides

Definition of compound Poisson processes

A compound Poisson process tracks the cumulative effect of random events that arrive according to a Poisson process, where each event carries a random "size" or "impact." Think of an insurance company: claims arrive randomly over time, and each claim has a different dollar amount. The compound Poisson process gives you the running total.

Formally, let N(t)N(t) be a Poisson process with rate λ\lambda, and let X1,X2,X_1, X_2, \ldots be independent and identically distributed (i.i.d.) random variables, also independent of N(t)N(t). The compound Poisson process is:

S(t)=i=1N(t)XiS(t) = \sum_{i=1}^{N(t)} X_i

Here S(t)S(t) represents the total accumulated value of all events up to time tt. When N(t)=0N(t) = 0, the sum is zero by convention.

Poisson process for event arrivals

The event count N(t)N(t) is an ordinary (homogeneous) Poisson process:

  • Events occur independently of one another.
  • The rate λ\lambda (average number of events per unit time) is constant.
  • The number of events in any interval of length tt follows a Poisson distribution with parameter λt\lambda t.
  • Inter-arrival times are exponentially distributed with parameter λ\lambda, so the expected time between events is 1λ\frac{1}{\lambda}.

Independent and identically distributed jump sizes

Each event ii carries a random variable XiX_i representing its magnitude. These jump sizes must satisfy two conditions:

  • Identically distributed: every XiX_i is drawn from the same distribution FXF_X.
  • Independent: the XiX_i are mutually independent and independent of the arrival process N(t)N(t).

The distribution FXF_X can be anything appropriate for the application: exponential for claim sizes, gamma for service times, lognormal for financial losses, etc. The independence between N(t)N(t) and the XiX_i is what makes the compound Poisson process analytically tractable.

Properties of compound Poisson processes

Moment generating function and PGF

The key analytical tool is the moment generating function (MGF). If MX(θ)=E[eθX]M_X(\theta) = E[e^{\theta X}] is the MGF of each jump XiX_i, then the MGF of S(t)S(t) has a clean closed form:

MS(t)(θ)=E[eθS(t)]=exp ⁣(λt(MX(θ)1))M_{S(t)}(\theta) = E[e^{\theta S(t)}] = \exp\!\bigl(\lambda t \bigl(M_X(\theta) - 1\bigr)\bigr)

This follows from conditioning on N(t)N(t) and using the Poisson PGF. For discrete-valued jumps, you can equivalently work with the probability generating function (PGF):

GS(t)(z)=GN(t) ⁣(GX(z))=exp ⁣(λt(GX(z)1))G_{S(t)}(z) = G_{N(t)}\!\bigl(G_X(z)\bigr) = \exp\!\bigl(\lambda t \bigl(G_X(z) - 1\bigr)\bigr)

Both forms let you extract moments by differentiation and identify the distribution of S(t)S(t) in many cases.

Moments of compound Poisson processes

Let μX=E[Xi]\mu_X = E[X_i] and σX2=Var(Xi)\sigma_X^2 = \text{Var}(X_i). Using the law of total expectation and total variance (conditioning on N(t)N(t)):

  • Mean: E[S(t)]=λtμXE[S(t)] = \lambda t \, \mu_X
  • Variance: Var(S(t))=λtE[Xi2]=λt(σX2+μX2)\text{Var}(S(t)) = \lambda t \, E[X_i^2] = \lambda t \,(\sigma_X^2 + \mu_X^2)

The variance formula deserves a closer look. It comes from the Eve's law (law of total variance) decomposition:

Var(S(t))=E[Var(SN)]+Var(E[SN])=λtσX2+λtμX2\text{Var}(S(t)) = E[\text{Var}(S \mid N)] + \text{Var}(E[S \mid N]) = \lambda t \, \sigma_X^2 + \lambda t \, \mu_X^2

The first term captures randomness in jump sizes; the second captures randomness in the number of jumps.

Stationary and independent increments

The compound Poisson process inherits the stationary and independent increments property from the underlying Poisson process. Concretely:

  • For any s>0s > 0, the increment S(t+s)S(t)S(t+s) - S(t) has the same distribution as S(s)S(s).
  • Increments over non-overlapping time intervals are independent.
  • Given the current value S(t)S(t), future increments don't depend on the path before time tt.

This is sometimes loosely called the "memoryless property," though that term more precisely refers to the exponential distribution. What matters here is that the process "resets" statistically at every point in time, which greatly simplifies calculations.

Examples of compound Poisson processes

Aggregate claims in insurance

An insurance company receives claims at rate λ=10\lambda = 10 per month. Each claim amount XiX_i follows a lognormal distribution with mean μX=$5,000\mu_X = \$5{,}000 and variance σX2\sigma_X^2. The total claims paid out by month tt is S(t)=i=1N(t)XiS(t) = \sum_{i=1}^{N(t)} X_i. The expected total payout over one month is E[S(1)]=10×5,000=$50,000E[S(1)] = 10 \times 5{,}000 = \$50{,}000. This model is foundational for premium setting and reserve calculations.

Cumulative damage models

In reliability engineering, a machine experiences random shocks at rate λ\lambda, and each shock causes damage XiX_i (perhaps Weibull- or exponentially distributed). The system fails when S(t)S(t) exceeds a threshold DD. The compound Poisson framework lets you compute the distribution of time to failure and optimize maintenance schedules.

Inventory demand modeling

Customer orders arrive at rate λ\lambda, and each order requests a random quantity XiX_i. The compound Poisson process S(t)S(t) gives total demand over [0,t][0, t], which is useful for setting reorder points and estimating stock-out probabilities. This model is especially appropriate for low-frequency, high-volume order patterns, like spare parts or wholesale goods.

Generalizations of compound Poisson processes

Poisson process for event arrivals, Statistics/Distributions/Poisson - Wikibooks, open books for an open world

Marked Poisson processes

A marked Poisson process attaches a random "mark" (label or attribute) to each event. The mark could encode event type, severity, location, or any other characteristic. A compound Poisson process is actually a special case: the mark is the jump size XiX_i, and you sum the marks. In the general marked process, you might analyze the marks without summing them.

Compound Cox processes

A Cox process (or doubly stochastic Poisson process) replaces the constant rate λ\lambda with a random intensity process Λ(t)\Lambda(t). A compound Cox process then sums i.i.d. jumps over this random-rate arrival process. This is useful when the event rate itself fluctuates unpredictably, as in financial markets (where trading intensity varies) or epidemiology (where infection rates change over time).

Compound renewal processes

A compound renewal process replaces the exponential inter-arrival times of the Poisson process with a general distribution (gamma, Weibull, lognormal, etc.). You lose the convenient Poisson structure and independent increments, but you gain flexibility to model arrivals where the exponential assumption is unrealistic. Analysis typically relies on renewal theory rather than the Poisson MGF formulas.

Simulation of compound Poisson processes

Simulating a compound Poisson process on [0,T][0, T] is straightforward. Here's the procedure:

  1. Generate the number of events. Draw NPoisson(λT)N \sim \text{Poisson}(\lambda T). This gives the total event count on [0,T][0, T].
  2. Generate event times. Draw NN uniform random variables U1,,UNU_1, \ldots, U_N on [0,T][0, T] and sort them. These are the arrival times. (Alternatively, generate exponential inter-arrival times and take cumulative sums, stopping when you exceed TT.)
  3. Generate jump sizes. For each event i=1,,Ni = 1, \ldots, N, sample XiX_i from the jump size distribution FXF_X using inverse transform sampling, acceptance-rejection, or a built-in generator.
  4. Compute the process. The value of S(t)S(t) at any time tt is the cumulative sum of all XiX_i whose arrival times fall in [0,t][0, t].

Repeating this procedure many times gives you Monte Carlo samples of S(T)S(T), from which you can estimate means, variances, tail probabilities, and other quantities of interest.

Parameter estimation for compound Poisson processes

Given observed data (event times and jump sizes), you need to estimate the rate λ\lambda and the parameters of the jump size distribution FXF_X.

Method of moments estimation

Match theoretical moments to sample moments:

  1. Estimate λ\lambda from the observed number of events per unit time: λ^=total eventsobservation period\hat{\lambda} = \frac{\text{total events}}{\text{observation period}}.
  2. Estimate μX\mu_X and σX2\sigma_X^2 from the sample mean and variance of the observed jump sizes.
  3. If jump sizes aren't directly observed (only S(t)S(t) increments are), use E[S(t)]=λtμXE[S(t)] = \lambda t \mu_X and Var(S(t))=λt(σX2+μX2)\text{Var}(S(t)) = \lambda t (\sigma_X^2 + \mu_X^2) to solve for the unknowns.

This approach is simple and fast but can be statistically inefficient, especially with small samples.

Maximum likelihood estimation

MLE constructs the likelihood from the observed data and maximizes it:

  • If both arrival times and jump sizes are observed, the likelihood factors cleanly: a Poisson process likelihood for the arrivals times a product of fX(xi)f_X(x_i) for the jumps. You can estimate λ\lambda and the jump distribution parameters separately.
  • If only aggregate increments S(tk+1)S(tk)S(t_{k+1}) - S(t_k) are observed, the likelihood involves the distribution of compound Poisson increments, which often lacks a closed form. Numerical optimization or the FFT-based approach to computing the compound distribution is then needed.

MLE estimators are asymptotically efficient and consistent, but may require iterative numerical methods.

Bayesian inference approaches

Bayesian estimation places prior distributions on λ\lambda and the jump size parameters, then updates them with observed data via Bayes' theorem:

  • A common choice is a Gamma prior for λ\lambda (conjugate to the Poisson likelihood).
  • Priors for jump size parameters depend on the assumed family FXF_X.
  • The posterior is computed analytically (in conjugate cases) or via MCMC sampling.

Bayesian methods naturally quantify parameter uncertainty through the posterior distribution and allow you to incorporate domain expertise through informative priors.

Applications of compound Poisson processes

Risk theory and ruin probabilities

The classical Cramér-Lundberg model describes an insurer's surplus as:

U(t)=u+ctS(t)U(t) = u + ct - S(t)

where uu is initial capital, cc is the premium income rate, and S(t)S(t) is the compound Poisson claim process. Ruin occurs if U(t)<0U(t) < 0 for some t>0t > 0. The ruin probability depends on the relationship between premium income and expected claims, and the tail behavior of the claim size distribution. For exponentially distributed claims with mean μX\mu_X, the ruin probability has the explicit form ψ(u)=λμXcexp ⁣(cλμXcμXu)\psi(u) = \frac{\lambda \mu_X}{c} \exp\!\left(-\frac{c - \lambda \mu_X}{c \mu_X} u\right), provided c>λμXc > \lambda \mu_X (the net profit condition).

Reliability analysis and shock models

Systems subject to random shocks degrade according to S(t)S(t). The system fails at the first time S(t)S(t) exceeds a damage threshold DD. Compound Poisson models let you compute the distribution of time-to-failure, optimize inspection intervals, and compare maintenance policies (e.g., age-based vs. condition-based replacement). These models appear in aerospace, power systems, and manufacturing.

Inventory management and demand modeling

With compound Poisson demand, you can derive the distribution of total demand over a lead time, which directly feeds into reorder point and safety stock calculations. For example, if orders arrive at rate λ=3\lambda = 3 per week with mean order size μX=50\mu_X = 50 units, expected weekly demand is 150150 units, and the variance of weekly demand is 3(σX2+2500)3(\sigma_X^2 + 2500). This variance drives safety stock decisions.