Binomial Distribution
Binomial distributions model experiments where you run a fixed number of trials and each trial has exactly two possible outcomes. They show up constantly in probability problems: coin flips, defective products on an assembly line, free-throw shooting percentages, survey responses. Understanding this distribution gives you a reliable framework for calculating how likely a particular number of "successes" is across repeated trials.

Binomial Distribution

Characteristics of binomial experiments
A binomial experiment has four specific conditions that must all be met. If even one is missing, you can't use the binomial model.
- Fixed number of trials (): The number of trials is determined before the experiment begins and doesn't change. You flip a coin 20 times, not "until you get bored."
- Two outcomes per trial: Each trial results in either a success () or a failure (). These labels are flexible: "success" just means the outcome you're tracking, even if it's something undesirable like a defective part.
- Constant probability of success (): The probability stays the same from one trial to the next. If on the first trial, it's on every trial.
- Independence between trials: The outcome of one trial doesn't affect the probability of success on any other trial. This is naturally satisfied with replacement sampling, or when the population is large enough relative to the sample that removing one item barely changes the probabilities (a common rule of thumb is the population should be at least 10 times the sample size).
A series of independent trials with a fixed probability of success is called Bernoulli trials. The binomial distribution counts the total number of successes across Bernoulli trials.

Formulas for binomial distribution statistics
Mean:
This gives the expected number of successes. If you flip a fair coin 50 times (, ), you'd expect heads.
Variance:
Variance measures how spread out the distribution of successes is around the mean. Notice that is just the probability of failure, sometimes written as .
Standard deviation:
This is the square root of the variance. It tells you, roughly, how far the actual number of successes typically falls from the expected value, in the same units as the count of successes.
Interpretation of binomial distribution measures
Mean (): This is the long-run average number of successes you'd see if you repeated the entire experiment many times. If and , then . That doesn't mean you'll get exactly 40 successes every time, but 40 is the center of the distribution.
Standard deviation (): This tells you how much the actual count of successes tends to vary from the mean.
- Using the same example (, ): . So in a typical run, you'd expect the number of successes to land within roughly 5 of the mean.
- A smaller means results cluster tightly around . A larger means more spread.
- is largest when (maximum uncertainty per trial) and shrinks as moves toward 0 or 1.
Probability functions and theorems
Probability mass function (PMF): This gives the probability of getting exactly successes in trials:
- is the binomial coefficient (" choose "), which counts the number of ways to arrange successes among trials.
- is the probability of successes occurring.
- is the probability of the remaining trials being failures.
For example, the probability of getting exactly 3 heads in 5 fair coin flips: .
Cumulative distribution function (CDF): This gives the probability of getting or fewer successes: . You calculate it by summing the PMF values from 0 through . Most calculators and tables handle this directly.
Connection to other theorems:
- The law of large numbers says that as you repeat the experiment more and more times, the observed proportion of successes converges toward .
- The central limit theorem tells you that when is large enough (commonly when both and ), the binomial distribution is well-approximated by a normal distribution with mean and standard deviation . This is useful because normal distribution calculations are often easier to perform.