Why This Matters
Cumulative Distribution Functions (CDFs) are one of the most versatile tools in probability. They show up everywhere from calculating probabilities to generating random samples in simulations. When you're tested on probability distributions, you're really being tested on your ability to move between CDFs, PDFs, and probability calculations.
The core idea: a CDF tracks how probability accumulates as you move along the number line. Don't just memorize that F(x) goes from 0 to 1. Understand why it must be non-decreasing (probability can't "un-accumulate") and how the CDF-PDF relationship lets you convert between cumulative and instantaneous probability descriptions. Those conceptual links are what application problems actually test.
Foundational Definitions and Properties
You need a rock-solid understanding of what a CDF is and what mathematical properties it must satisfy. These properties aren't arbitrary; they follow directly from the axioms of probability.
Definition of Cumulative Distribution Function
F(x)=P(Xโคx)
This is the probability that random variable X takes a value less than or equal to x. Two things to note:
- Complete distribution description. Knowing the CDF tells you everything about the random variable's probabilistic behavior. You can recover the PDF, compute any probability, find percentiles, etc.
- Range is always [0,1]. Since F(x) is a probability, it can never go below 0 or above 1.
Properties of CDFs
Every valid CDF must satisfy three properties:
- Non-decreasing. If x1โ<x2โ, then F(x1โ)โคF(x2โ). This is because the event {Xโคx1โ} is a subset of {Xโคx2โ}, so its probability can't be larger.
- Boundary limits. limxโโโโF(x)=0 and limxโโโF(x)=1. Far enough to the left, no probability has accumulated yet; far enough to the right, all of it has.
- Right-continuous. F(x)=limhโ0+โF(x+h). This convention ensures that F(x) includes the probability at exactly x. It matters most for discrete distributions, where the CDF has jump discontinuities.
Compare: Non-decreasing vs. Right-continuous. Both are required CDF properties, but non-decreasing reflects probability accumulation while right-continuity is a technical convention about how jumps are handled. Exam problems often test whether you can spot an invalid CDF that violates one of these.
The CDF-PDF Relationship
The PDF tells you probability density at a point; the CDF tells you accumulated probability up to that point. Knowing how to go back and forth between them is essential.
Relationship for Continuous Variables
- PDF is the derivative of CDF: f(x)=dxdF(x)โ. The PDF measures the rate at which probability accumulates.
- CDF is the integral of PDF: F(x)=โซโโxโf(t)dt. This gives the area under the PDF curve from โโ to x.
- Probability over an interval is the area under f(x) between the endpoints: P(a<Xโคb)=โซabโf(t)dt=F(b)โF(a).
Continuous vs. Discrete CDFs
- Continuous CDFs are smooth curves with no jumps. You can differentiate them to get the PDF.
- Discrete CDFs are step functions. Each jump occurs at a possible outcome xiโ, and the jump height equals the PMF value P(X=xiโ). You use differences (not derivatives) to recover probabilities.
- Both types completely describe their distribution. The difference is how probability accumulates: smoothly vs. in discrete chunks.
Compare: If a problem gives you a step function and asks for probabilities, you're working with a discrete distribution. Use differences: P(X=xiโ)=F(xiโ)โF(xiโโ), where F(xiโโ) is the CDF value just before the jump. If the CDF is smooth, differentiate to get the PDF.
Probability Calculations with CDFs
This is where CDFs prove their practical value. They make probability calculations straightforward, especially for ranges and tail probabilities.
Calculating Probabilities Using CDFs
The key formulas:
- Range probability: P(a<Xโคb)=F(b)โF(a). This works for both continuous and discrete variables.
- Tail probability: P(X>x)=1โF(x). You'll use this constantly in reliability and survival analysis contexts.
- For continuous variables specifically, P(X=x)=0 for any single point, so P(a<Xโคb)=P(aโคXโคb). The inequality direction doesn't matter. For discrete variables, it does matter whether the endpoints are included.
Inverse CDF and Its Applications
The quantile function Fโ1(p) goes the other direction: given a probability p, it returns the value x such that P(Xโคx)=p.
- Finding percentiles. The median is Fโ1(0.5). The 95th percentile is Fโ1(0.95).
- Random sample generation. The inverse transform method works like this: generate a uniform random number U on [0,1], then compute X=Fโ1(U). The result X follows the target distribution. This works for any distribution with an invertible CDF.
- Confidence intervals. Critical values for hypothesis testing come from the inverse CDF of the test statistic's distribution.
Compare: CDF vs. Inverse CDF. The CDF maps values to probabilities (xโp). The inverse CDF maps probabilities to values (pโx). Simulation problems typically need the inverse CDF; probability calculations typically need the CDF directly.
CDFs for Standard Distributions
Knowing the CDF shapes and formulas for common distributions lets you quickly identify distribution types and apply the right methods.
CDFs for Common Probability Distributions
Normal distribution. The CDF is the S-shaped curve written as ฮฆ(z) for the standard normal. There's no closed-form expression, so you look up values in a table or use software. The symmetry property ฮฆ(โz)=1โฮฆ(z) is useful for quick calculations.
Exponential distribution. F(x)=1โeโฮปx for xโฅ0. Probability accumulates quickly at first, then slows. This distribution has the memoryless property: P(X>s+tโฃX>s)=P(X>t). The parameter ฮป is the rate; the mean is 1/ฮป.
Uniform distribution. F(x)=bโaxโaโ for xโ[a,b]. This is just a straight line from 0 to 1, reflecting the constant density across the interval.
Extensions and Advanced Applications
These topics extend CDF concepts to multiple variables and real-world data analysis.
Joint CDFs for Multiple Random Variables
The joint CDF for two random variables is defined as:
F(x,y)=P(Xโคx,Yโคy)
This describes the simultaneous behavior of X and Y. It extends naturally to n variables: F(x1โ,x2โ,โฆ,xnโ).
A key use: checking independence. If X and Y are independent, then F(x,y)=FXโ(x)โ
FYโ(y) for all x,y. If this product relationship fails anywhere, the variables are dependent.
Empirical CDF and Data Analysis
When you have actual data (say n observations), the empirical CDF is:
F^(x)=n1โโi=1nโ1(Xiโโคx)
In plain terms, F^(x) is the fraction of your data points that fall at or below x. This is a non-parametric estimate, meaning you don't need to assume any particular distribution.
The empirical CDF is especially useful for goodness-of-fit testing. The Kolmogorov-Smirnov test compares the empirical CDF to a theoretical CDF and measures the maximum vertical distance between them. A large gap suggests the data doesn't follow the hypothesized distribution.
Compare: Theoretical vs. Empirical CDF. Theoretical CDFs come from assumed distributions with known parameters. Empirical CDFs are built directly from data. Use empirical CDFs when you don't know the true distribution or want to validate model assumptions.
- Probability integral transform. If X has a continuous CDF F, then F(X) follows a Uniform(0,1) distribution. This is the theoretical foundation behind the inverse transform method for simulation.
- Standardization. Transforming data using CDFs can map any continuous distribution to a uniform one, which is useful for comparing distributions or meeting statistical assumptions.
Quick Reference Table
|
| Core definition | F(x)=P(Xโคx), non-decreasing, right-continuous, boundary limits 0 and 1 |
| CDF-PDF relationship | f(x)=dxdFโ and F(x)=โซโโxโf(t)dt |
| Probability calculations | P(a<Xโคb)=F(b)โF(a), tail: P(X>x)=1โF(x) |
| Inverse CDF | Quantile function Fโ1(p), used for percentiles and simulation |
| Continuous vs. discrete | Smooth curves vs. step functions; derivatives vs. differences |
| Common distributions | Normal (ฮฆ(z), S-curve), Exponential (1โeโฮปx), Uniform (linear) |
| Joint CDFs | F(x,y)=P(Xโคx,Yโคy); independence โบ F(x,y)=FXโ(x)FYโ(y) |
| Empirical CDF | F^(x)=n#{Xiโโคx}โ, used in Kolmogorov-Smirnov test |
Self-Check Questions
-
A function G(x) satisfies G(โโ)=0 and G(โ)=1, but decreases on some interval. Can G(x) be a valid CDF? Why or why not?
-
Compare how you would find P(2<Xโค5) using a CDF versus using a PDF. When might one approach be preferable?
-
You need to generate random samples from an Exponential distribution with rate ฮป. Describe how the inverse CDF method works and write the transformation formula.
-
Given a dataset of 100 observations, explain how you would construct the empirical CDF and what information it gives you that a histogram doesn't.
-
If two random variables X and Y are independent, what relationship must hold between their joint CDF F(x,y) and their marginal CDFs FXโ(x) and FYโ(y)?