Fiveable

📊Actuarial Mathematics Unit 1 Review

QR code for Actuarial Mathematics practice questions

1.3 Random variables and probability distributions

1.3 Random variables and probability distributions

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📊Actuarial Mathematics
Unit & Topic Study Guides

Random variables and probability distributions are the foundation of actuarial mathematics. They give you the tools to assign numbers to uncertain outcomes and then describe how likely those outcomes are, which is exactly what actuaries need to quantify risk in insurance and finance.

This section covers the types of random variables, how probability distributions work (including joint, marginal, and conditional distributions), the most common discrete and continuous distributions, moments, transformations, and limit theorems.

Types of random variables

A random variable assigns a numerical value to each outcome of a random experiment. The type of random variable you're working with determines which mathematical tools you'll use, so getting this distinction right matters.

Discrete vs continuous

Discrete random variables take on countable values, like integers or elements of a finite set. Think of things you can count: the number of claims filed in a month, the number of policies sold, or the number of defaults in a portfolio.

Continuous random variables can take any value within an interval. These describe things you measure rather than count: the dollar amount of a claim, the time until a policy lapses, or an insured's age at death.

The distinction matters because discrete and continuous variables use different mathematical machinery (summation vs. integration), and mixing them up will lead to errors.

Probability mass functions

A probability mass function (PMF) describes the probability distribution of a discrete random variable. It tells you the probability that the variable equals each possible value.

  • Notation: P(X=x)P(X = x), where XX is the random variable and xx is a specific value
  • Every probability must be non-negative, and the sum across all possible values must equal 1: xP(X=x)=1\sum_x P(X = x) = 1
  • The binomial and Poisson distributions are common examples of distributions defined by PMFs

Probability density functions

A probability density function (PDF) describes the distribution of a continuous random variable. Unlike a PMF, a PDF does not give you the probability of a single point. Instead, it gives the relative likelihood at each value.

  • Notation: f(x)f(x)
  • To find the probability that XX falls in an interval, you integrate: P(aXb)=abf(x)dxP(a \leq X \leq b) = \int_a^b f(x)\, dx
  • The total area under the PDF must equal 1: f(x)dx=1\int_{-\infty}^{\infty} f(x)\, dx = 1
  • The normal and exponential distributions are common examples

A key point that trips people up: for a continuous random variable, P(X=x)=0P(X = x) = 0 for any single value xx. Probabilities only make sense over intervals.

Probability distributions

Probability distributions are mathematical functions that describe how likely different outcomes are for a random variable. Actuaries rely on them to model everything from claim frequencies to loss severities.

Cumulative distribution functions

The cumulative distribution function (CDF) gives the probability that a random variable is less than or equal to a specific value:

F(x)=P(Xx)F(x) = P(X \leq x)

CDFs work for both discrete and continuous random variables:

  • Discrete: F(x)=txP(X=t)F(x) = \sum_{t \leq x} P(X = t)
  • Continuous: F(x)=xf(t)dtF(x) = \int_{-\infty}^{x} f(t)\, dt

Every CDF has three properties: it's non-decreasing, limxF(x)=0\lim_{x \to -\infty} F(x) = 0, and limxF(x)=1\lim_{x \to \infty} F(x) = 1. CDFs are especially useful for computing probabilities over intervals and for finding quantiles (percentiles) of a distribution.

Joint probability distributions

Joint probability distributions describe the probability of two or more random variables taking on values simultaneously.

  • For discrete variables, the joint PMF is P(X=x,Y=y)P(X = x, Y = y)
  • For continuous variables, the joint PDF is f(x,y)f(x, y), and probabilities are found by integrating over a region

Joint distributions let you model dependent risks. For example, you might use a joint distribution to capture the relationship between claim frequency and claim severity for an insurance policy. The joint CDF is F(x,y)=P(Xx,Yy)F(x, y) = P(X \leq x, Y \leq y).

Marginal probability distributions

Marginal distributions are obtained from a joint distribution by "summing out" or "integrating out" the other variables:

  • Discrete: P(X=x)=yP(X=x,Y=y)P(X = x) = \sum_y P(X = x, Y = y)
  • Continuous: fX(x)=f(x,y)dyf_X(x) = \int_{-\infty}^{\infty} f(x, y)\, dy

The result is the distribution of a single variable on its own, ignoring the others. For instance, if you have the joint distribution of claim frequency and severity, the marginal distribution of frequency tells you about frequency alone.

Conditional probability distributions

Conditional distributions describe the probability of one variable given a known value of another:

  • Discrete: P(X=xY=y)=P(X=x,Y=y)P(Y=y)P(X = x \mid Y = y) = \frac{P(X = x, Y = y)}{P(Y = y)}
  • Continuous: f(xy)=f(x,y)fY(y)f(x \mid y) = \frac{f(x, y)}{f_Y(y)}

These are essential for updating your model when new information arrives. For example, the distribution of claim severity given that a claim has occurred is a conditional distribution. Conditional distributions connect to joint and marginal distributions through Bayes' theorem.

Common discrete distributions

Discrete distributions model counted quantities. Each distribution below has specific assumptions, so choosing the right one depends on the structure of the problem.

Bernoulli distribution

Models a single trial with two outcomes: success (1) or failure (0).

  • Notation: XBern(p)X \sim \text{Bern}(p), where pp is the probability of success
  • PMF: P(X=x)=px(1p)1xP(X = x) = p^x (1-p)^{1-x} for x{0,1}x \in \{0, 1\}
  • Mean: pp, Variance: p(1p)p(1-p)

Use this for binary events, such as whether a policyholder files a claim or not.

Binomial distribution

Models the number of successes in nn independent Bernoulli trials, each with the same success probability pp.

  • Notation: XBin(n,p)X \sim \text{Bin}(n, p)
  • PMF: P(X=x)=(nx)px(1p)nxP(X = x) = \binom{n}{x} p^x (1-p)^{n-x} for x{0,1,,n}x \in \{0, 1, \ldots, n\}
  • Mean: npnp, Variance: np(1p)np(1-p)

Use this when you have a fixed number of independent trials. For example, out of 100 policies, how many will result in a claim this year?

Discrete vs continuous, 4.1: Probability Distribution Function (PDF) for a Discrete Random Variable - Statistics LibreTexts

Poisson distribution

Models the count of events occurring in a fixed interval of time or space, where events happen independently at a constant average rate.

  • Notation: XPois(λ)X \sim \text{Pois}(\lambda), where λ\lambda is the average number of events per interval
  • PMF: P(X=x)=eλλxx!P(X = x) = \frac{e^{-\lambda}\lambda^x}{x!} for x{0,1,2,}x \in \{0, 1, 2, \ldots\}
  • Mean: λ\lambda, Variance: λ\lambda

The Poisson is the workhorse distribution for claim frequency modeling. A notable property: the mean equals the variance. If your data shows the variance significantly exceeding the mean, the Poisson may not be the right fit (consider the negative binomial instead).

Geometric distribution

Models the number of trials until the first success in a sequence of independent Bernoulli trials.

  • Notation: XGeom(p)X \sim \text{Geom}(p)
  • PMF: P(X=x)=(1p)x1pP(X = x) = (1-p)^{x-1}p for x{1,2,3,}x \in \{1, 2, 3, \ldots\}
  • Mean: 1/p1/p, Variance: (1p)/p2(1-p)/p^2

Watch out: some textbooks define the geometric as the number of failures before the first success, which shifts the support to {0,1,2,}\{0, 1, 2, \ldots\} and changes the PMF to P(X=x)=(1p)xpP(X = x) = (1-p)^x p. Always check which convention is being used.

Negative binomial distribution

Generalizes the geometric distribution. Models the number of failures before achieving rr successes.

  • Notation: XNB(r,p)X \sim \text{NB}(r, p)
  • PMF: P(X=x)=(x+r1x)(1p)xprP(X = x) = \binom{x+r-1}{x}(1-p)^x p^r for x{0,1,2,}x \in \{0, 1, 2, \ldots\}
  • Mean: r(1p)/pr(1-p)/p, Variance: r(1p)/p2r(1-p)/p^2

In actuarial work, the negative binomial is often used as an alternative to the Poisson for claim counts when overdispersion is present (i.e., the variance exceeds the mean).

Hypergeometric distribution

Models the number of successes in nn draws from a finite population of size NN containing KK successes, without replacement.

  • Notation: XHGeom(N,K,n)X \sim \text{HGeom}(N, K, n)
  • PMF: P(X=x)=(Kx)(NKnx)(Nn)P(X = x) = \frac{\binom{K}{x}\binom{N-K}{n-x}}{\binom{N}{n}} for x{max(0,n+KN),,min(n,K)}x \in \{\max(0, n+K-N), \ldots, \min(n, K)\}

The key difference from the binomial: sampling is done without replacement, so trials are not independent. Use this when the population is small enough that removing items changes the probabilities meaningfully. As NN grows large relative to nn, the hypergeometric approaches the binomial.

Common continuous distributions

Continuous distributions model measured quantities like claim amounts, durations, and financial returns.

Uniform distribution

Every value in the interval [a,b][a, b] is equally likely.

  • Notation: XUnif(a,b)X \sim \text{Unif}(a, b)
  • PDF: f(x)=1baf(x) = \frac{1}{b-a} for x[a,b]x \in [a, b]
  • Mean: (a+b)/2(a+b)/2, Variance: (ba)2/12(b-a)^2/12

Often used when you have no prior information about which values are more likely, or as a building block in simulation (since many random number generators produce uniform variates that get transformed into other distributions).

Normal distribution

The symmetric, bell-shaped distribution that appears throughout statistics and actuarial science, largely because of the central limit theorem.

  • Notation: XN(μ,σ2)X \sim \mathcal{N}(\mu, \sigma^2), where μ\mu is the mean and σ2\sigma^2 is the variance
  • PDF: f(x)=12πσ2e(xμ)22σ2f(x) = \frac{1}{\sqrt{2\pi\sigma^2}}e^{-\frac{(x-\mu)^2}{2\sigma^2}} for xRx \in \mathbb{R}

The standard normal has μ=0\mu = 0 and σ2=1\sigma^2 = 1. Any normal variable can be standardized: Z=(Xμ)/σZ = (X - \mu)/\sigma. Used to model aggregate losses (via CLT), financial returns, and as an approximation for many other distributions when sample sizes are large.

Exponential distribution

Models the waiting time until the next event in a Poisson process.

  • Notation: XExp(λ)X \sim \text{Exp}(\lambda), where λ\lambda is the rate parameter
  • PDF: f(x)=λeλxf(x) = \lambda e^{-\lambda x} for x0x \geq 0
  • Mean: 1/λ1/\lambda, Variance: 1/λ21/\lambda^2

The exponential distribution has the memoryless property: P(X>s+tX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t). This means the probability of waiting an additional tt units doesn't depend on how long you've already waited. It's the only continuous distribution with this property.

Gamma distribution

Generalizes the exponential. When α\alpha is a positive integer, it models the waiting time until the α\alpha-th event in a Poisson process.

  • Notation: XGamma(α,β)X \sim \text{Gamma}(\alpha, \beta), where α\alpha is the shape and β\beta is the rate
  • PDF: f(x)=βαΓ(α)xα1eβxf(x) = \frac{\beta^\alpha}{\Gamma(\alpha)}x^{\alpha-1}e^{-\beta x} for x>0x > 0
  • Mean: α/β\alpha/\beta, Variance: α/β2\alpha/\beta^2

Note that Γ(α)\Gamma(\alpha) is the gamma function, which generalizes the factorial: Γ(n)=(n1)!\Gamma(n) = (n-1)! for positive integers. The gamma distribution is widely used for modeling claim amounts and in credibility theory. When α=1\alpha = 1, it reduces to the exponential.

Beta distribution

Models a random variable constrained to the interval (0,1)(0, 1), making it natural for proportions and probabilities.

  • Notation: XBeta(α,β)X \sim \text{Beta}(\alpha, \beta)
  • PDF: f(x)=1B(α,β)xα1(1x)β1f(x) = \frac{1}{B(\alpha, \beta)}x^{\alpha-1}(1-x)^{\beta-1} for x(0,1)x \in (0, 1)
  • Mean: α/(α+β)\alpha/(\alpha + \beta), Variance: αβ/[(α+β)2(α+β+1)]\alpha\beta/[(\alpha+\beta)^2(\alpha+\beta+1)]

Here B(α,β)=Γ(α)Γ(β)/Γ(α+β)B(\alpha, \beta) = \Gamma(\alpha)\Gamma(\beta)/\Gamma(\alpha+\beta) is the beta function. The beta distribution is very flexible: depending on the parameters, it can be symmetric, left-skewed, right-skewed, U-shaped, or uniform (when α=β=1\alpha = \beta = 1). Actuaries use it for loss ratios and as a prior distribution in Bayesian analysis.

Lognormal distribution

If ln(X)\ln(X) follows a normal distribution, then XX follows a lognormal distribution. This means XX is always positive and right-skewed.

  • Notation: XLognormal(μ,σ2)X \sim \text{Lognormal}(\mu, \sigma^2), where μ\mu and σ2\sigma^2 are the mean and variance of ln(X)\ln(X)
  • PDF: f(x)=1x2πσ2e(lnxμ)22σ2f(x) = \frac{1}{x\sqrt{2\pi\sigma^2}}e^{-\frac{(\ln x - \mu)^2}{2\sigma^2}} for x>0x > 0
  • Mean: eμ+σ2/2e^{\mu + \sigma^2/2}, Variance: (eσ21)e2μ+σ2(e^{\sigma^2} - 1)e^{2\mu + \sigma^2}

The lognormal is one of the most common distributions for modeling individual claim amounts because claims are positive and their distribution is typically right-skewed (many small claims, few very large ones).

Discrete vs continuous, Discrete Random Variables (3 of 5) | Concepts in Statistics

Moments of distributions

Moments are numerical summaries that capture the shape and behavior of a distribution. The first few moments tell you about the center, spread, asymmetry, and tail behavior.

Expected value

The expected value (mean) is the probability-weighted average of all possible values.

  • Notation: E[X]\mathbb{E}[X] or μ\mu
  • Discrete: E[X]=xxP(X=x)\mathbb{E}[X] = \sum_{x} x\, P(X = x)
  • Continuous: E[X]=xf(x)dx\mathbb{E}[X] = \int_{-\infty}^{\infty} x\, f(x)\, dx

The expected value is linear: E[aX+b]=aE[X]+b\mathbb{E}[aX + b] = a\mathbb{E}[X] + b. Actuaries use it to calculate fair premiums, expected claim costs, and expected profits.

Variance and standard deviation

Variance measures how spread out the distribution is around the mean.

  • Notation: Var(X)\text{Var}(X) or σ2\sigma^2
  • Computed as: Var(X)=E[(Xμ)2]=E[X2](E[X])2\text{Var}(X) = \mathbb{E}[(X - \mu)^2] = \mathbb{E}[X^2] - (\mathbb{E}[X])^2

The second formula (E[X2](E[X])2\mathbb{E}[X^2] - (\mathbb{E}[X])^2) is usually easier to compute. The standard deviation σ=Var(X)\sigma = \sqrt{\text{Var}(X)} has the same units as XX, making it more interpretable.

Variance is used to assess volatility of claim amounts, set risk margins, and determine capital requirements.

Skewness and kurtosis

Skewness measures asymmetry:

γ1=E[(Xμσ)3]\gamma_1 = \mathbb{E}\left[\left(\frac{X - \mu}{\sigma}\right)^3\right]

  • γ1>0\gamma_1 > 0: right-skewed (long right tail, common for claim amounts)
  • γ1<0\gamma_1 < 0: left-skewed
  • γ1=0\gamma_1 = 0: symmetric

Excess kurtosis measures tail heaviness relative to the normal distribution:

γ2=E[(Xμσ)4]3\gamma_2 = \mathbb{E}\left[\left(\frac{X - \mu}{\sigma}\right)^4\right] - 3

The "3-3" makes the normal distribution the baseline (γ2=0\gamma_2 = 0). Positive excess kurtosis means heavier tails than normal, which signals a higher probability of extreme outcomes. This is critical for actuaries assessing catastrophic risk.

Moment-generating functions

The moment-generating function (MGF) encodes all the moments of a distribution in a single function:

MX(t)=E[etX]M_X(t) = \mathbb{E}[e^{tX}]

To extract the nn-th moment, take the nn-th derivative and evaluate at t=0t = 0:

E[Xn]=MX(n)(0)\mathbb{E}[X^n] = M_X^{(n)}(0)

MGFs are powerful for three reasons:

  1. If two random variables have the same MGF (and it exists in a neighborhood of 0), they have the same distribution
  2. For independent random variables, MX+Y(t)=MX(t)MY(t)M_{X+Y}(t) = M_X(t) \cdot M_Y(t), which simplifies working with sums
  3. They provide a clean way to prove limit theorems

Not every distribution has an MGF (the Cauchy distribution, for example, does not). When the MGF doesn't exist, the characteristic function serves as an alternative.

Transformations of random variables

Transformations create new random variables by applying functions to existing ones. This is how you move between distributions and model more complex relationships.

Linear transformations

A linear transformation takes the form Y=aX+bY = aX + b, where aa and bb are constants.

  • Mean: E[Y]=aE[X]+b\mathbb{E}[Y] = a\mathbb{E}[X] + b
  • Variance: Var(Y)=a2Var(X)\text{Var}(Y) = a^2\text{Var}(X)

Notice that adding a constant bb shifts the mean but doesn't affect the variance. Multiplying by aa scales both the mean and the standard deviation (by a|a|), and scales the variance by a2a^2.

Common uses: standardizing a variable (Z=(Xμ)/σZ = (X - \mu)/\sigma), converting units, or adjusting claim amounts for inflation.

Functions of random variables

For a general transformation Y=g(X)Y = g(X), you need to find the distribution of YY.

The CDF method works as follows:

  1. Start with FY(y)=P(g(X)y)F_Y(y) = P(g(X) \leq y)

  2. Rewrite the event {g(X)y}\{g(X) \leq y\} in terms of XX

  3. If gg is monotonically increasing: FY(y)=FX(g1(y))F_Y(y) = F_X(g^{-1}(y))

  4. If gg is monotonically decreasing: FY(y)=1FX(g1(y))F_Y(y) = 1 - F_X(g^{-1}(y))

  5. Differentiate the CDF to get the PDF of YY

For a strictly monotonic, differentiable gg, the PDF can be found directly:

fY(y)=fX(g1(y))ddyg1(y)f_Y(y) = f_X(g^{-1}(y)) \cdot \left|\frac{d}{dy}g^{-1}(y)\right|

This technique is how you derive, for example, that exponentiating a normal variable gives a lognormal variable.

Convolutions and sums of random variables

When XX and YY are independent, the distribution of Z=X+YZ = X + Y is found through convolution:

  • Discrete: P(Z=z)=xP(X=x)P(Y=zx)P(Z = z) = \sum_{x} P(X = x)\,P(Y = z - x)
  • Continuous: fZ(z)=fX(x)fY(zx)dxf_Z(z) = \int_{-\infty}^{\infty} f_X(x)\,f_Y(z - x)\, dx

In practice, convolution integrals can be difficult to compute directly. The MGF approach is often easier: since MX+Y(t)=MX(t)MY(t)M_{X+Y}(t) = M_X(t) \cdot M_Y(t) for independent variables, you can multiply the MGFs and then identify the resulting distribution.

For example, the sum of independent Poisson random variables with parameters λ1\lambda_1 and λ2\lambda_2 is Poisson with parameter λ1+λ2\lambda_1 + \lambda_2. You can verify this quickly using MGFs.

Convolutions are central to modeling aggregate losses, where total claims equal the sum of individual claim amounts.

Limit theorems

Limit theorems describe what happens as you work with larger and larger samples. They justify many of the approximations actuaries use in practice.

Law of large numbers

The law of large numbers (LLN) says that the sample mean converges to the population mean as the sample size grows.

For i.i.d. random variables X1,X2,X_1, X_2, \ldots with mean μ\mu:

Xˉn=1ni=1nXiμas n\bar{X}_n = \frac{1}{n}\sum_{i=1}^n X_i \to \mu \quad \text{as } n \to \infty

There are two versions:

  • Weak LLN: Xˉn\bar{X}_n converges to μ\mu in probability
  • Strong LLN: Xˉn\bar{X}_n converges to μ\mu almost surely (a stronger guarantee)

The LLN is the theoretical foundation of insurance: with a large enough pool of policyholders, the average claim per policy becomes predictable. This is why insurers can set stable premiums despite individual claims being random.

Central limit theorem

The central limit theorem (CLT) states that the standardized sum (or average) of a large number of i.i.d. random variables is approximately normally distributed, regardless of the original distribution.

For i.i.d. random variables with mean μ\mu and variance σ2\sigma^2:

Xˉnμσ/ndN(0,1)as n\frac{\bar{X}_n - \mu}{\sigma / \sqrt{n}} \xrightarrow{d} \mathcal{N}(0, 1) \quad \text{as } n \to \infty

Equivalently, for the sum Sn=i=1nXiS_n = \sum_{i=1}^n X_i:

SnnμσndN(0,1)\frac{S_n - n\mu}{\sigma\sqrt{n}} \xrightarrow{d} \mathcal{N}(0, 1)

The CLT is why the normal distribution appears so frequently in practice. Actuaries use it to approximate the distribution of aggregate claims, construct confidence intervals, and perform hypothesis tests. As a rough guideline, the approximation tends to work well for n30n \geq 30, though highly skewed distributions may require larger samples.