Geometric Distribution
Geometric distributions model the number of failures before a first success in repeated independent trials. They come up whenever you're asking "how many attempts until something happens?" Whether that's flipping a coin until you get heads, rolling a die until you land a six, or testing products until one fails inspection, the geometric distribution gives you the framework to calculate probabilities, expected values, and variability for these situations.

Characteristics of Geometric Experiments
A geometric experiment is built from Bernoulli trials, which are independent trials that each have exactly two outcomes: success or failure. What makes it geometric is that you keep running trials until you get your first success, then stop.
Four conditions must hold for a geometric distribution to apply:
- Independence: The outcome of one trial doesn't affect any other. A coin doesn't "remember" its last flip.
- Constant probability: The probability of success (and failure ) stays the same on every trial. For a fair coin, every time. For rolling a six on a fair die, every time.
- Two outcomes per trial: Each trial results in either success or failure.
- The experiment ends at the first success: You stop as soon as success occurs.
The geometric distribution also has the memoryless property. This means that no matter how many failures you've already observed, the probability of success on the very next trial is still . If you've flipped 10 tails in a row, the probability of heads on flip 11 is still 0.5. Past failures give you no information about when the next success will come.

Probability Calculation in Geometric Distributions
The probability mass function (PMF) tells you the probability of observing exactly failures before the first success:
where is the random variable counting failures before the first success, is a specific number of failures (0, 1, 2, ...), and is the probability of success on each trial.
How to use the PMF:
- Identify (the probability of success) and (the number of failures you're interested in).
- Plug into the formula.
- Compute.
Example: What's the probability of getting exactly 3 tails before your first heads on a fair coin?
Here and :
There's a 6.25% chance of exactly 3 tails before the first heads.
Watch the definition of X. Some textbooks define as the number of trials until the first success (so includes the success trial itself), which shifts the formula to . Check which version your course uses. This guide uses = number of failures before the first success.
The cumulative distribution function (CDF) gives the probability of observing at most failures before the first success:
This is useful when you need probabilities for ranges rather than a single value. For instance, the probability of getting your first heads within the first 4 flips (i.e., 0, 1, 2, or 3 failures) on a fair coin is .

Analysis of Geometric Probability Distributions
The mean (expected value) tells you the average number of failures you'd expect before the first success:
The variance measures how spread out the number of failures tends to be:
The standard deviation is the square root of the variance:
Example: For a fair coin ():
- failure on average before the first heads
So on average you'd expect 1 tail before your first heads, with a standard deviation of about 1.41.
Example with a rarer event: For rolling a six on a fair die ():
- failures on average before the first six
Notice how a smaller leads to both a higher expected number of failures and much more variability. When success is rare, the distribution spreads out considerably.
Related Distributions and Concepts
The negative binomial distribution generalizes the geometric distribution. Instead of counting failures before the first success, it counts failures before the -th success. The geometric distribution is the special case where .
The probabilities in a geometric distribution form a geometric sequence (each term is a constant multiple of the previous one), which is where the distribution gets its name. This connection to geometric series is also why the probabilities sum to 1: the infinite sum converges to 1 for any .