Fiveable

🎲Intro to Statistics Unit 4 Review

QR code for Intro to Statistics practice questions

4.2 Mean or Expected Value and Standard Deviation

4.2 Mean or Expected Value and Standard Deviation

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
🎲Intro to Statistics
Unit & Topic Study Guides

Expected Value and Standard Deviation

Expected value and standard deviation give you two essential pieces of information about any discrete random variable: where the center is and how spread out the values are. These tools let you summarize an entire probability distribution with just two numbers, which is the basis for making predictions and comparing outcomes.

Mean of Discrete Probability Distributions

The mean (or expected value) of a discrete random variable is the long-run average you'd expect after repeating an experiment many, many times. It's denoted by μ\mu or E(X)E(X).

You calculate it by multiplying each possible outcome by its probability, then adding all those products together:

E(X)=μ=i=1nxiP(X=xi)E(X) = \mu = \sum_{i=1}^{n} x_i \cdot P(X = x_i)

  • xix_i = each possible value the random variable can take
  • P(X=xi)P(X = x_i) = the probability of that value occurring
  • nn = the total number of possible values

Example: Rolling a fair die. Each face (1 through 6) has probability 16\frac{1}{6}, so:

E(X)=1(16)+2(16)+3(16)+4(16)+5(16)+6(16)=3.5E(X) = 1\left(\frac{1}{6}\right) + 2\left(\frac{1}{6}\right) + 3\left(\frac{1}{6}\right) + 4\left(\frac{1}{6}\right) + 5\left(\frac{1}{6}\right) + 6\left(\frac{1}{6}\right) = 3.5

Notice that 3.5 isn't a value you can actually roll. That's fine. The expected value doesn't have to be a possible outcome; it represents the theoretical average over many rolls.

Mean of discrete probability distributions, Discrete Random Variables (5 of 5) | Concepts in Statistics

Standard Deviation of Discrete Distributions

The standard deviation (σ\sigma) tells you how far values typically fall from the mean. A small standard deviation means outcomes cluster tightly around μ\mu; a large one means they're more spread out.

To get there, you first calculate the variance (σ2\sigma^2):

Var(X)=σ2=i=1n(xiμ)2P(X=xi)Var(X) = \sigma^2 = \sum_{i=1}^{n} (x_i - \mu)^2 \cdot P(X = x_i)

Then the standard deviation is the square root of the variance: σ=σ2\sigma = \sqrt{\sigma^2}.

Steps to calculate standard deviation:

  1. Compute the mean μ\mu using the expected value formula.

  2. For each possible value xix_i, subtract the mean: (xiμ)(x_i - \mu).

  3. Square each of those differences: (xiμ)2(x_i - \mu)^2.

  4. Multiply each squared difference by its probability: (xiμ)2P(X=xi)(x_i - \mu)^2 \cdot P(X = x_i).

  5. Add up all those products. This sum is the variance.

  6. Take the square root of the variance to get the standard deviation.

Example: Suppose a random variable XX has this distribution:

xx135
P(X=x)P(X = x)0.20.50.3
  • Mean: μ=1(0.2)+3(0.5)+5(0.3)=0.2+1.5+1.5=3.2\mu = 1(0.2) + 3(0.5) + 5(0.3) = 0.2 + 1.5 + 1.5 = 3.2
  • Variance: (13.2)2(0.2)+(33.2)2(0.5)+(53.2)2(0.3)=(4.84)(0.2)+(0.04)(0.5)+(3.24)(0.3)=0.968+0.02+0.972=1.96(1-3.2)^2(0.2) + (3-3.2)^2(0.5) + (5-3.2)^2(0.3) = (4.84)(0.2) + (0.04)(0.5) + (3.24)(0.3) = 0.968 + 0.02 + 0.972 = 1.96
  • Standard deviation: σ=1.96=1.4\sigma = \sqrt{1.96} = 1.4

A common mistake is forgetting to weight by the probabilities. You're not just averaging the squared differences; you're taking a probability-weighted average.

Mean of discrete probability distributions, Average - Wikipedia

Law of Large Numbers Interpretation

The Law of Large Numbers is what connects expected value to the real world. It says that as you repeat an experiment more and more times, the sample mean (the actual average of your results) gets closer and closer to the expected value μ\mu.

This also applies to relative frequencies. As the number of trials grows, the relative frequency of an event approaches its true probability:

Relative frequency of A=Number of times A occursTotal number of trials\text{Relative frequency of A} = \frac{\text{Number of times A occurs}}{\text{Total number of trials}}

Example: Flip a fair coin 10 times and you might get 7 heads (relative frequency = 0.70). Flip it 10,000 times and the relative frequency of heads will be very close to 0.50. The expected value doesn't guarantee any single result; it describes what happens in the long run.

This is why expected value matters: it's not a prediction for one trial, but a reliable summary of what happens over many trials.