Expected Value and Standard Deviation
Expected value and standard deviation give you two essential pieces of information about any discrete random variable: where the center is and how spread out the values are. These tools let you summarize an entire probability distribution with just two numbers, which is the basis for making predictions and comparing outcomes.
Mean of Discrete Probability Distributions
The mean (or expected value) of a discrete random variable is the long-run average you'd expect after repeating an experiment many, many times. It's denoted by or .
You calculate it by multiplying each possible outcome by its probability, then adding all those products together:
- = each possible value the random variable can take
- = the probability of that value occurring
- = the total number of possible values
Example: Rolling a fair die. Each face (1 through 6) has probability , so:
Notice that 3.5 isn't a value you can actually roll. That's fine. The expected value doesn't have to be a possible outcome; it represents the theoretical average over many rolls.

Standard Deviation of Discrete Distributions
The standard deviation () tells you how far values typically fall from the mean. A small standard deviation means outcomes cluster tightly around ; a large one means they're more spread out.
To get there, you first calculate the variance ():
Then the standard deviation is the square root of the variance: .
Steps to calculate standard deviation:
-
Compute the mean using the expected value formula.
-
For each possible value , subtract the mean: .
-
Square each of those differences: .
-
Multiply each squared difference by its probability: .
-
Add up all those products. This sum is the variance.
-
Take the square root of the variance to get the standard deviation.
Example: Suppose a random variable has this distribution:
| 1 | 3 | 5 | |
|---|---|---|---|
| 0.2 | 0.5 | 0.3 |
- Mean:
- Variance:
- Standard deviation:
A common mistake is forgetting to weight by the probabilities. You're not just averaging the squared differences; you're taking a probability-weighted average.

Law of Large Numbers Interpretation
The Law of Large Numbers is what connects expected value to the real world. It says that as you repeat an experiment more and more times, the sample mean (the actual average of your results) gets closer and closer to the expected value .
This also applies to relative frequencies. As the number of trials grows, the relative frequency of an event approaches its true probability:
Example: Flip a fair coin 10 times and you might get 7 heads (relative frequency = 0.70). Flip it 10,000 times and the relative frequency of heads will be very close to 0.50. The expected value doesn't guarantee any single result; it describes what happens in the long run.
This is why expected value matters: it's not a prediction for one trial, but a reliable summary of what happens over many trials.