Fiveable

📊Honors Statistics Unit 7 Review

QR code for Honors Statistics practice questions

7.2 The Central Limit Theorem for Sums (Optional)

7.2 The Central Limit Theorem for Sums (Optional)

Written by the Fiveable Content Team • Last updated August 2025
Written by the Fiveable Content Team • Last updated August 2025
📊Honors Statistics
Unit & Topic Study Guides
Pep mascot

The Central Limit Theorem for Sums

The Central Limit Theorem (CLT) for sums extends the same core idea you learned for sample means: when you add up a large number of independent random variables, their sum follows an approximately normal distribution, no matter what the original population looks like. This matters because it lets you use normal distribution techniques (z-scores, probability tables) to answer questions about sums, even when the underlying data isn't normal.

Pep mascot
more resources to help you study

The Central Limit Theorem for Sums

The CLT for sums says that if you take a large enough sample of independent, identically distributed (i.i.d.) random variables and add them up, the distribution of that sum will be approximately normal. The key requirement is that the sample size nn is sufficiently large. The more skewed or unusual the population distribution, the larger nn needs to be for the approximation to work well.

Compare this to the CLT for means: both theorems rely on the same convergence principle, but here you're looking at the total rather than the average.

Parameters of the sampling distribution for sums:

  • Mean: μX=nμ\mu_{\sum X} = n\mu
  • Standard deviation: σX=nσ\sigma_{\sum X} = \sqrt{n} \cdot \sigma

where nn is the sample size, μ\mu is the population mean, and σ\sigma is the population standard deviation.

Notice that the mean of the sum scales linearly with nn, but the standard deviation scales with n\sqrt{n}. This is why variability grows more slowly than the total itself, and it's the reason sums become more predictable (relative to their size) as nn increases.

Mean and Standard Deviation of the Sampling Distribution

To find the center and spread of the sampling distribution for sums, plug into the two formulas above.

Example: Suppose a population has μ=10\mu = 10 and σ=2\sigma = 2, and you draw a sample of n=25n = 25.

  1. Calculate the mean of the sum: μX=25×10=250\mu_{\sum X} = 25 \times 10 = 250
  2. Calculate the standard deviation of the sum: σX=25×2=5×2=10\sigma_{\sum X} = \sqrt{25} \times 2 = 5 \times 2 = 10

So the sampling distribution for the sum of 25 observations is approximately normal with a center at 250 and a standard deviation of 10.

Central Limit Theorem for sums, Introduction to Normal Random Variables | Concepts in Statistics

Z-Scores for Sums

Once you know the sampling distribution is approximately normal, you can standardize any observed sum using:

z=XμXσXz = \frac{\sum X - \mu_{\sum X}}{\sigma_{\sum X}}

where X\sum X is the observed sum, μX\mu_{\sum X} is the mean of the sampling distribution for sums, and σX\sigma_{\sum X} is its standard deviation. The z-score tells you how many standard deviations a particular sum falls from the expected total.

Step-by-step process for finding a probability:

  1. Identify nn, μ\mu, and σ\sigma from the problem.
  2. Compute μX=nμ\mu_{\sum X} = n\mu and σX=nσ\sigma_{\sum X} = \sqrt{n} \cdot \sigma.
  3. Plug the observed (or target) sum into the z-score formula.
  4. Use the standard normal table (or calculator) to find the corresponding probability.

Example: Using the parameters above (μX=250\mu_{\sum X} = 250, σX=10\sigma_{\sum X} = 10), what's the probability the sum exceeds 265?

  1. z=26525010=1.5z = \frac{265 - 250}{10} = 1.5

  2. From the z-table, P(Z<1.5)=0.9332P(Z < 1.5) = 0.9332

  3. So P(X>265)=10.9332=0.0668P(\sum X > 265) = 1 - 0.9332 = 0.0668, or about 6.7%.

You can also work backward from a probability to find a percentile. For instance, to find the 90th percentile of the sum, look up z=1.28z = 1.28 and solve X=μX+zσX=250+1.28(10)=262.8\sum X = \mu_{\sum X} + z \cdot \sigma_{\sum X} = 250 + 1.28(10) = 262.8.

Statistical Foundations and Applications

A few related ideas tie into this theorem:

  • Law of Large Numbers: As nn grows, the sample mean converges to the population mean. The CLT for sums builds on this by describing how the sum's distribution behaves, not just where it centers.
  • Statistical inference: The CLT for sums is what justifies using normal-based methods (confidence intervals, hypothesis tests) for totals. Without it, you'd need to know the exact population distribution to calculate probabilities for sums.
  • The CLT for sums and the CLT for means are directly connected. Since Xˉ=Xn\bar{X} = \frac{\sum X}{n}, any result about the sum can be translated into a result about the mean, and vice versa. They're two sides of the same theorem.