upgrade
upgrade

🎲Intro to Probability

Expected Value Formulas

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

Expected value is the backbone of probability theory—it's how we quantify what "should" happen on average when randomness is involved. Whether you're analyzing casino games, insurance policies, or scientific experiments, expected value gives you the single number that summarizes a random variable's long-run behavior. You're being tested on your ability to calculate expected values across different distribution types, apply key properties like linearity and conditional expectation, and recognize which formula fits which scenario.

Don't just memorize these formulas in isolation. The real exam skill is understanding when to use each one and why it works. Can you recognize a binomial setup and immediately write down E(X)=npE(X) = np? Do you know why linearity of expectation is so powerful? Focus on connecting each formula to the underlying principle it represents—that's what separates a 5 from a 3.


Core Definitions: Discrete vs. Continuous

The fundamental expected value formulas differ based on whether your random variable takes countable values or spans a continuous range. The discrete case uses summation; the continuous case uses integration.

Expected Value of a Discrete Random Variable

  • E(X)=xiP(xi)E(X) = \sum x_i \cdot P(x_i)—multiply each possible value by its probability, then sum all products
  • Weighted average interpretation—values with higher probabilities contribute more to the expected value
  • Center of distribution—represents where the probability "mass" balances, useful for comparing random variables

Expected Value of a Continuous Random Variable

  • E(X)=xf(x)dxE(X) = \int_{-\infty}^{\infty} x \cdot f(x) \, dx—where f(x)f(x) is the probability density function (PDF)
  • Balance point concept—if you cut out the PDF shape from cardboard, the expected value is where it would balance
  • Analogous to discrete case—integration replaces summation when dealing with uncountably many possible values

Compare: Discrete vs. Continuous EV—both calculate a weighted average, but discrete uses \sum while continuous uses \int. If an FRQ gives you a PDF, you're integrating; if it gives you a probability table, you're summing.


Key Properties: Simplifying Calculations

These properties are your computational shortcuts. Linearity especially shows up constantly because it works even when variables aren't independent.

Linearity of Expectation

  • E(X+Y)=E(X)+E(Y)E(X + Y) = E(X) + E(Y)—always true, regardless of independence
  • Extends to constantsE(aX+b)=aE(X)+bE(aX + b) = aE(X) + b for any constants aa and bb
  • Problem-solving power—break complex random variables into simpler pieces, find each EV, then add

Expected Value of a Function of a Random Variable

  • E(g(X))=g(xi)P(xi)E(g(X)) = \sum g(x_i) P(x_i) for discrete; E(g(X))=g(x)f(x)dxE(g(X)) = \int g(x) f(x) \, dx for continuous
  • LOTUS (Law of the Unconscious Statistician)—you don't need the distribution of g(X)g(X), just apply gg inside the expectation
  • Common application—finding E(X2)E(X^2) to calculate variance via Var(X)=E(X2)[E(X)]2\text{Var}(X) = E(X^2) - [E(X)]^2

Compare: Linearity vs. LOTUS—linearity handles sums of random variables; LOTUS handles functions of a single random variable. Both simplify calculations, but they solve different problems.


Conditional Expectation: Incorporating Information

When you know something about one variable, it changes what you expect from another. Conditional expectation formalizes how new information updates your predictions.

Conditional Expected Value

  • E(XY=y)=xiP(X=xiY=y)E(X | Y = y) = \sum x_i P(X = x_i | Y = y) for discrete; uses conditional PDF for continuous
  • Updates the average—given that YY takes a specific value, what's the new "best guess" for XX?
  • Measures dependence—if E(XY)E(X|Y) varies with YY, then XX and YY are dependent

Law of Total Expectation

  • E(X)=E(E(XY))E(X) = E(E(X | Y))—the overall expected value equals the average of conditional expected values
  • Partition strategy—break a problem into cases based on YY, solve each case, then recombine
  • Also called iterated expectation—essential for multi-stage problems like "first flip a coin, then roll dice based on the result"

Compare: Conditional EV vs. Law of Total Expectation—conditional EV gives you the answer for a specific scenario; the law of total expectation averages across all scenarios. Use the law when you need the unconditional answer but the conditional calculation is easier.


Named Distributions: Memorize These Formulas

For common distributions, the expected value formulas are derived once and used forever. Know these cold—they're tested directly and save enormous time.

Expected Value of a Binomial Distribution

  • E(X)=npE(X) = np—where nn is the number of trials and pp is the probability of success per trial
  • Intuitive meaning—in 100 coin flips with p=0.5p = 0.5, expect 50 heads on average
  • Derived from linearity—a binomial is the sum of nn independent Bernoulli trials, each with E=pE = p

Expected Value of a Poisson Distribution

  • E(X)=λE(X) = \lambda—where λ\lambda is the average rate of occurrence
  • Parameter does double duty—for Poisson, the mean equals the variance equals λ\lambda
  • Models rare events—number of emails per hour, accidents per month, typos per page

Compare: Binomial vs. Poisson—binomial has a fixed number of trials; Poisson models events in a continuous interval. As nn \to \infty and p0p \to 0 with np=λnp = \lambda, binomial approaches Poisson.

Expected Value of an Exponential Distribution

  • E(X)=1λE(X) = \frac{1}{\lambda}—where λ\lambda is the rate parameter
  • Waiting time interpretation—average time until the next event in a Poisson process
  • Memoryless property connection—the exponential distribution "forgets" how long you've already waited

Expected Value of a Uniform Distribution

  • E(X)=a+b2E(X) = \frac{a + b}{2}—simply the midpoint of the interval [a,b][a, b]
  • Symmetry argument—all values equally likely, so the average is the center
  • Quick sanity check—if your uniform is on [0,10][0, 10], expected value must be 5

Compare: Exponential vs. Uniform—both are continuous, but exponential is skewed right (most values near zero) while uniform is symmetric. Their expected value formulas reflect this: exponential depends only on rate, uniform depends only on boundaries.


Quick Reference Table

ConceptBest Examples
Discrete EV formulaE(X)=xiP(xi)E(X) = \sum x_i P(x_i)
Continuous EV formulaE(X)=xf(x)dxE(X) = \int x f(x) dx
Linearity propertyE(X+Y)=E(X)+E(Y)E(X + Y) = E(X) + E(Y), E(aX+b)=aE(X)+bE(aX + b) = aE(X) + b
Function of RV (LOTUS)E(g(X))E(g(X)) using original distribution
Conditional expectationE(XY)E(X \mid Y), Law of Total Expectation
BinomialE(X)=npE(X) = np
PoissonE(X)=λE(X) = \lambda
ExponentialE(X)=1/λE(X) = 1/\lambda
UniformE(X)=(a+b)/2E(X) = (a+b)/2

Self-Check Questions

  1. What property allows you to calculate E(X1+X2++Xn)E(X_1 + X_2 + \cdots + X_n) without knowing whether the variables are independent?

  2. Compare the expected value formulas for Poisson and exponential distributions—how are their parameters related, and why does this make sense given the Poisson process interpretation?

  3. If XX is uniform on [2,8][2, 8] and YY is binomial with n=12n = 12 and p=0.5p = 0.5, which has the larger expected value? Show your work.

  4. Explain when you would use the Law of Total Expectation instead of calculating E(X)E(X) directly. Give an example scenario.

  5. You need to find E(X2)E(X^2) for a discrete random variable. Would you use linearity of expectation or LOTUS? Write the formula you'd apply.