๐ŸƒEngineering Probability

Cumulative Distribution Functions

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

Cumulative Distribution Functions (CDFs) are one of the most versatile tools in probability. They show up everywhere from calculating probabilities to generating random samples in simulations. When you're tested on probability distributions, you're really being tested on your ability to move between CDFs, PDFs, and probability calculations.

The core idea: a CDF tracks how probability accumulates as you move along the number line. Don't just memorize that F(x)F(x) goes from 0 to 1. Understand why it must be non-decreasing (probability can't "un-accumulate") and how the CDF-PDF relationship lets you convert between cumulative and instantaneous probability descriptions. Those conceptual links are what application problems actually test.


Foundational Definitions and Properties

You need a rock-solid understanding of what a CDF is and what mathematical properties it must satisfy. These properties aren't arbitrary; they follow directly from the axioms of probability.

Definition of Cumulative Distribution Function

F(x)=P(Xโ‰คx)F(x) = P(X \leq x)

This is the probability that random variable XX takes a value less than or equal to xx. Two things to note:

  • Complete distribution description. Knowing the CDF tells you everything about the random variable's probabilistic behavior. You can recover the PDF, compute any probability, find percentiles, etc.
  • Range is always [0,1][0, 1]. Since F(x)F(x) is a probability, it can never go below 0 or above 1.

Properties of CDFs

Every valid CDF must satisfy three properties:

  1. Non-decreasing. If x1<x2x_1 < x_2, then F(x1)โ‰คF(x2)F(x_1) \leq F(x_2). This is because the event {Xโ‰คx1}\{X \leq x_1\} is a subset of {Xโ‰คx2}\{X \leq x_2\}, so its probability can't be larger.
  2. Boundary limits. limโกxโ†’โˆ’โˆžF(x)=0\lim_{x \to -\infty} F(x) = 0 and limโกxโ†’โˆžF(x)=1\lim_{x \to \infty} F(x) = 1. Far enough to the left, no probability has accumulated yet; far enough to the right, all of it has.
  3. Right-continuous. F(x)=limโกhโ†’0+F(x+h)F(x) = \lim_{h \to 0^+} F(x + h). This convention ensures that F(x)F(x) includes the probability at exactly xx. It matters most for discrete distributions, where the CDF has jump discontinuities.

Compare: Non-decreasing vs. Right-continuous. Both are required CDF properties, but non-decreasing reflects probability accumulation while right-continuity is a technical convention about how jumps are handled. Exam problems often test whether you can spot an invalid CDF that violates one of these.


The CDF-PDF Relationship

The PDF tells you probability density at a point; the CDF tells you accumulated probability up to that point. Knowing how to go back and forth between them is essential.

Relationship for Continuous Variables

  • PDF is the derivative of CDF: f(x)=dF(x)dxf(x) = \frac{dF(x)}{dx}. The PDF measures the rate at which probability accumulates.
  • CDF is the integral of PDF: F(x)=โˆซโˆ’โˆžxf(t)โ€‰dtF(x) = \int_{-\infty}^{x} f(t) \, dt. This gives the area under the PDF curve from โˆ’โˆž-\infty to xx.
  • Probability over an interval is the area under f(x)f(x) between the endpoints: P(a<Xโ‰คb)=โˆซabf(t)โ€‰dt=F(b)โˆ’F(a)P(a < X \leq b) = \int_a^b f(t) \, dt = F(b) - F(a).

Continuous vs. Discrete CDFs

  • Continuous CDFs are smooth curves with no jumps. You can differentiate them to get the PDF.
  • Discrete CDFs are step functions. Each jump occurs at a possible outcome xix_i, and the jump height equals the PMF value P(X=xi)P(X = x_i). You use differences (not derivatives) to recover probabilities.
  • Both types completely describe their distribution. The difference is how probability accumulates: smoothly vs. in discrete chunks.

Compare: If a problem gives you a step function and asks for probabilities, you're working with a discrete distribution. Use differences: P(X=xi)=F(xi)โˆ’F(xiโˆ’)P(X = x_i) = F(x_i) - F(x_i^-), where F(xiโˆ’)F(x_i^-) is the CDF value just before the jump. If the CDF is smooth, differentiate to get the PDF.


Probability Calculations with CDFs

This is where CDFs prove their practical value. They make probability calculations straightforward, especially for ranges and tail probabilities.

Calculating Probabilities Using CDFs

The key formulas:

  • Range probability: P(a<Xโ‰คb)=F(b)โˆ’F(a)P(a < X \leq b) = F(b) - F(a). This works for both continuous and discrete variables.
  • Tail probability: P(X>x)=1โˆ’F(x)P(X > x) = 1 - F(x). You'll use this constantly in reliability and survival analysis contexts.
  • For continuous variables specifically, P(X=x)=0P(X = x) = 0 for any single point, so P(a<Xโ‰คb)=P(aโ‰คXโ‰คb)P(a < X \leq b) = P(a \leq X \leq b). The inequality direction doesn't matter. For discrete variables, it does matter whether the endpoints are included.

Inverse CDF and Its Applications

The quantile function Fโˆ’1(p)F^{-1}(p) goes the other direction: given a probability pp, it returns the value xx such that P(Xโ‰คx)=pP(X \leq x) = p.

  • Finding percentiles. The median is Fโˆ’1(0.5)F^{-1}(0.5). The 95th percentile is Fโˆ’1(0.95)F^{-1}(0.95).
  • Random sample generation. The inverse transform method works like this: generate a uniform random number UU on [0,1][0, 1], then compute X=Fโˆ’1(U)X = F^{-1}(U). The result XX follows the target distribution. This works for any distribution with an invertible CDF.
  • Confidence intervals. Critical values for hypothesis testing come from the inverse CDF of the test statistic's distribution.

Compare: CDF vs. Inverse CDF. The CDF maps values to probabilities (xโ†’px \to p). The inverse CDF maps probabilities to values (pโ†’xp \to x). Simulation problems typically need the inverse CDF; probability calculations typically need the CDF directly.


CDFs for Standard Distributions

Knowing the CDF shapes and formulas for common distributions lets you quickly identify distribution types and apply the right methods.

CDFs for Common Probability Distributions

Normal distribution. The CDF is the S-shaped curve written as ฮฆ(z)\Phi(z) for the standard normal. There's no closed-form expression, so you look up values in a table or use software. The symmetry property ฮฆ(โˆ’z)=1โˆ’ฮฆ(z)\Phi(-z) = 1 - \Phi(z) is useful for quick calculations.

Exponential distribution. F(x)=1โˆ’eโˆ’ฮปxF(x) = 1 - e^{-\lambda x} for xโ‰ฅ0x \geq 0. Probability accumulates quickly at first, then slows. This distribution has the memoryless property: P(X>s+tโˆฃX>s)=P(X>t)P(X > s + t \mid X > s) = P(X > t). The parameter ฮป\lambda is the rate; the mean is 1/ฮป1/\lambda.

Uniform distribution. F(x)=xโˆ’abโˆ’aF(x) = \frac{x - a}{b - a} for xโˆˆ[a,b]x \in [a, b]. This is just a straight line from 0 to 1, reflecting the constant density across the interval.


Extensions and Advanced Applications

These topics extend CDF concepts to multiple variables and real-world data analysis.

Joint CDFs for Multiple Random Variables

The joint CDF for two random variables is defined as:

F(x,y)=P(Xโ‰คx,Yโ‰คy)F(x, y) = P(X \leq x, Y \leq y)

This describes the simultaneous behavior of XX and YY. It extends naturally to nn variables: F(x1,x2,โ€ฆ,xn)F(x_1, x_2, \ldots, x_n).

A key use: checking independence. If XX and YY are independent, then F(x,y)=FX(x)โ‹…FY(y)F(x, y) = F_X(x) \cdot F_Y(y) for all x,yx, y. If this product relationship fails anywhere, the variables are dependent.

Empirical CDF and Data Analysis

When you have actual data (say nn observations), the empirical CDF is:

F^(x)=1nโˆ‘i=1n1(Xiโ‰คx)\hat{F}(x) = \frac{1}{n} \sum_{i=1}^{n} \mathbf{1}(X_i \leq x)

In plain terms, F^(x)\hat{F}(x) is the fraction of your data points that fall at or below xx. This is a non-parametric estimate, meaning you don't need to assume any particular distribution.

The empirical CDF is especially useful for goodness-of-fit testing. The Kolmogorov-Smirnov test compares the empirical CDF to a theoretical CDF and measures the maximum vertical distance between them. A large gap suggests the data doesn't follow the hypothesized distribution.

Compare: Theoretical vs. Empirical CDF. Theoretical CDFs come from assumed distributions with known parameters. Empirical CDFs are built directly from data. Use empirical CDFs when you don't know the true distribution or want to validate model assumptions.

CDF Transformations and Applications

  • Probability integral transform. If XX has a continuous CDF FF, then F(X)F(X) follows a Uniform(0,1)\text{Uniform}(0, 1) distribution. This is the theoretical foundation behind the inverse transform method for simulation.
  • Standardization. Transforming data using CDFs can map any continuous distribution to a uniform one, which is useful for comparing distributions or meeting statistical assumptions.

Quick Reference Table

ConceptKey Facts
Core definitionF(x)=P(Xโ‰คx)F(x) = P(X \leq x), non-decreasing, right-continuous, boundary limits 0 and 1
CDF-PDF relationshipf(x)=dFdxf(x) = \frac{dF}{dx} and F(x)=โˆซโˆ’โˆžxf(t)โ€‰dtF(x) = \int_{-\infty}^{x} f(t)\,dt
Probability calculationsP(a<Xโ‰คb)=F(b)โˆ’F(a)P(a < X \leq b) = F(b) - F(a), tail: P(X>x)=1โˆ’F(x)P(X > x) = 1 - F(x)
Inverse CDFQuantile function Fโˆ’1(p)F^{-1}(p), used for percentiles and simulation
Continuous vs. discreteSmooth curves vs. step functions; derivatives vs. differences
Common distributionsNormal (ฮฆ(z)\Phi(z), S-curve), Exponential (1โˆ’eโˆ’ฮปx1 - e^{-\lambda x}), Uniform (linear)
Joint CDFsF(x,y)=P(Xโ‰คx,Yโ‰คy)F(x,y) = P(X \leq x, Y \leq y); independence โ€…โ€ŠโŸบโ€…โ€Š\iff F(x,y)=FX(x)FY(y)F(x,y) = F_X(x) F_Y(y)
Empirical CDFF^(x)=#{Xiโ‰คx}n\hat{F}(x) = \frac{\#\{X_i \leq x\}}{n}, used in Kolmogorov-Smirnov test

Self-Check Questions

  1. A function G(x)G(x) satisfies G(โˆ’โˆž)=0G(-\infty) = 0 and G(โˆž)=1G(\infty) = 1, but decreases on some interval. Can G(x)G(x) be a valid CDF? Why or why not?

  2. Compare how you would find P(2<Xโ‰ค5)P(2 < X \leq 5) using a CDF versus using a PDF. When might one approach be preferable?

  3. You need to generate random samples from an Exponential distribution with rate ฮป\lambda. Describe how the inverse CDF method works and write the transformation formula.

  4. Given a dataset of 100 observations, explain how you would construct the empirical CDF and what information it gives you that a histogram doesn't.

  5. If two random variables XX and YY are independent, what relationship must hold between their joint CDF F(x,y)F(x, y) and their marginal CDFs FX(x)F_X(x) and FY(y)F_Y(y)?