Mixture models
Mixture models let you represent a population as a combination of distinct subpopulations, each with its own probability distribution. This is critical in actuarial work because real insurance portfolios are rarely homogeneous: your book of business might contain low-risk drivers, moderate-risk drivers, and high-risk drivers, each generating claims from a different underlying distribution. A single distribution can't capture that kind of structure, but a mixture model can.
Definition of mixture models
A mixture model expresses the overall density of a random variable as a weighted sum of component distributions. Each component represents a subpopulation, and the weights reflect how prevalent each subpopulation is. The result is a single model flexible enough to capture multimodal behavior, heavy tails, or other patterns that no single standard distribution handles well.
Components of mixture models
A mixture model has two building blocks:
- Mixing proportions : the probability that a randomly chosen observation comes from component . These must satisfy and each .
- Component distributions : the PDF (or PMF) for each subpopulation, parameterized by . Common choices include exponential, gamma, lognormal, Pareto, and normal distributions.
The overall density is:
For example, a two-component mixture of exponentials with means 1,000 and 10,000 and mixing weights 0.7 and 0.3 would produce a distribution with a bulk of small claims and a meaningful tail of large claims.
Finite vs. infinite mixture models
- Finite mixture models fix the number of components in advance. You either choose from domain knowledge (e.g., "we know there are three risk classes") or select it using model selection criteria like AIC or BIC.
- Infinite mixture models (Dirichlet process mixtures) allow the number of components to grow with the data. These are useful when you have no strong prior belief about how many subpopulations exist, but they require Bayesian computational methods and are more complex to implement.
For most exam-level and practical actuarial work, you'll deal with finite mixtures.
Applications in actuarial science
- Insurance pricing: Segment policyholders into risk classes, each with its own claim distribution, to set appropriate premiums.
- Claims modeling: Represent small, medium, and large claims as separate components. A two-component mixture (e.g., lognormal + Pareto) is a classic approach for capturing both the body and the tail of a severity distribution.
- Reserving: Different claim types may develop at different speeds; mixture models can account for these distinct development patterns.
- Fraud detection and classification: Unusual claims may form their own component, making them identifiable through posterior component probabilities.
Deductibles
A deductible is the amount of loss a policyholder must pay out of pocket before the insurer begins to pay. Deductibles are a fundamental risk-sharing mechanism: they reduce insurer exposure, discourage frivolous claims, and lower premiums.
Purpose of deductibles
- Reduce moral hazard: When policyholders bear part of the loss, they have an incentive to take preventive measures.
- Eliminate small claims: Insurers avoid the administrative cost of processing many low-dollar claims.
- Lower premiums: Shifting some risk to the policyholder means the insurer charges less.
Types of deductibles
Ordinary (fixed) deductible: The policyholder pays the first dollars of any loss. The insurer pays , where is the loss amount. A $500 deductible on an $1,800 loss means the insurer pays $1,300.
Percentage deductible: Expressed as a percentage of the insured value. A 2% deductible on a $100,000 property means the policyholder absorbs the first $2,000.
Franchise deductible: If the loss is below , the policyholder pays everything. If the loss meets or exceeds , the insurer pays the entire loss. With a $1,000 franchise deductible, a $900 claim is fully on the policyholder, but a $1,500 claim is fully covered by the insurer. This differs sharply from an ordinary deductible, where the insurer would only pay $500 on that $1,500 claim.
Impact on insurance premiums
Higher deductibles transfer more risk to the policyholder, which reduces the insurer's expected claim payments and therefore lowers the premium. The relationship isn't linear: going from a $0 to $500 deductible typically produces a larger premium reduction than going from $500 to $1,000, because most of the claim frequency sits in the lower severity range.
Actuaries quantify this by computing the loss elimination ratio (LER):
This gives the proportion of expected losses eliminated by a deductible of .
Deductibles vs. coinsurance
Both are risk-sharing tools, but they operate differently:
- A deductible applies first: the policyholder absorbs losses up to .
- Coinsurance applies after the deductible: the policyholder pays a percentage of the remaining loss. With 80/20 coinsurance and a $500 deductible, on a $2,500 loss the insurer pays and the policyholder pays $900.
These mechanisms are often combined in a single policy.

Modeling claims with deductibles
When a deductible exists, the insurer only observes claims that exceed the deductible, and the amount paid differs from the ground-truth loss. You need to adjust the severity distribution accordingly. There are two key perspectives, and mixing them up is a common source of errors.
Claim size distribution
The ground-truth loss follows some severity distribution (exponential, gamma, Pareto, lognormal, etc.). The choice depends on the data's characteristics: Pareto or lognormal for heavy-tailed data, exponential or gamma for lighter tails.
Per-loss variable (left-truncated distribution)
This models the original loss given that it exceeds the deductible. You're conditioning on , so you use a left-truncated distribution:
Here is the original PDF and is the CDF at . The denominator rescales the density so it integrates to 1 over .
This is the distribution of losses given that a claim is reported. It's sometimes called the per-loss variable in exam notation.
Per-payment variable (shifted/excess loss distribution)
The insurer doesn't pay ; it pays for losses exceeding . This is the per-payment or excess loss variable. Its PDF is:
Notice this combines both the shift (replacing with ) and the truncation (dividing by ). The variable represents the actual dollar amount the insurer pays on each claim.
Be careful: the shifted distribution alone (without the denominator) does not integrate to 1 over unless . The correct per-payment density always includes the survival function in the denominator.
Mixture of truncated and shifted distributions
In some actuarial models, you combine these perspectives using a mixture. For instance, if a portfolio contains both claims that exceed the deductible and claims that are handled differently (e.g., partial payments or different coverage layers), you might write:
The mixing proportion represents the probability that a claim falls into the first category. This structure appears when modeling portfolios with heterogeneous deductible arrangements.
Parameter estimation
With mixture models, you're estimating both the mixing proportions and the parameters of each component distribution. This is harder than fitting a single distribution because you don't know which component generated each observation.
Maximum likelihood estimation
The log-likelihood for a -component mixture with observations is:
where .
The sum inside the logarithm makes direct optimization difficult. You can't simply take the derivative and solve in closed form. That's why the EM algorithm is the standard approach.
Expectation-maximization algorithm
The EM algorithm handles the "missing data" problem (you don't know which component each observation belongs to) by iterating between two steps:
- E-step: Compute the posterior probability that observation belongs to component , using current parameter estimates:
These values are called responsibilities: they tell you how much each component "claims" each data point.
- M-step: Update the parameters using the responsibilities as weights:
- Repeat until the log-likelihood converges (i.e., changes by less than some tolerance between iterations).
A practical warning: EM can converge to a local maximum, so it's standard practice to run it from multiple random starting points and keep the solution with the highest log-likelihood.
Bayesian inference for mixture models
Bayesian methods treat the parameters as random variables with prior distributions . After observing data , you update to the posterior:
Because the posterior for mixture models is analytically intractable, you sample from it using MCMC methods:
- Gibbs sampling works well when you can derive the full conditional distributions for each parameter block (common with conjugate priors).
- Metropolis-Hastings is more general but requires tuning proposal distributions.
Bayesian inference naturally quantifies parameter uncertainty through the posterior distribution and supports model comparison via Bayes factors.

Model selection
Choosing the right number of components is one of the most important decisions when fitting a mixture model. Too few components underfit; too many overfit.
Criteria for model selection
All standard criteria balance fit against complexity. The key trade-off: adding a component always improves the likelihood, but it also adds parameters that may just be fitting noise.
Akaike information criterion (AIC)
where is the total number of estimated parameters. Lower AIC is better. AIC tends to favor slightly more complex models and is useful when the goal is prediction rather than identifying the "true" model.
Bayesian information criterion (BIC)
where is the sample size. The penalty term grows with sample size, so BIC penalizes complexity more heavily than AIC for any . BIC is consistent, meaning it selects the true number of components as (assuming the true model is among the candidates). In practice, BIC tends to choose simpler models than AIC.
Likelihood ratio tests
For comparing nested models (e.g., vs. ), the likelihood ratio statistic is:
Under the null hypothesis, approximately follows a chi-squared distribution with degrees of freedom equal to the difference in parameter counts.
There's a subtlety here: standard regularity conditions for the chi-squared approximation can fail for mixture models because under the null, some parameters are on the boundary of the parameter space (e.g., ). Adjusted tests or bootstrap methods are sometimes needed.
Simulation of mixture models
Simulation lets you generate synthetic data from a fitted mixture, which is useful for validating estimation methods, stress testing, and computing quantities that are hard to derive analytically.
Generating random variates
To simulate one observation from a -component mixture:
- Draw a component label from a categorical distribution with probabilities .
- Given , draw from the component distribution using inverse transform sampling, acceptance-rejection, or a built-in generator.
- Repeat for as many observations as needed.
This two-step process directly mirrors the generative interpretation of the mixture model.
Monte Carlo simulation
Once you can generate data from the mixture, you can estimate any quantity by simulation:
- Moments: Compute the sample mean and variance of a large simulated sample to approximate and .
- Tail probabilities: Estimate as the fraction of simulated values exceeding .
- Quantiles: Sort the simulated values and read off the desired percentile (e.g., VaR at the 99th percentile).
The accuracy of Monte Carlo estimates improves with the number of simulations, roughly at a rate of where is the number of replications.
Assessing model fit
After fitting a mixture model, check whether it actually describes the data well:
- Histogram overlay: Plot the fitted mixture density on top of a histogram of the observed data. Gaps or systematic deviations signal a poor fit.
- Q-Q plot: Compare the quantiles of the observed data against the theoretical quantiles of the fitted mixture. Points should fall near the 45-degree line.
- Formal tests: The Kolmogorov-Smirnov test compares the empirical CDF to the fitted CDF. The chi-squared goodness-of-fit test bins the data and compares observed vs. expected frequencies. Both can flag significant departures from the model, though with large samples even minor deviations become "significant."