Comparing Two Population Means with Known Standard Deviations
When you want to test whether two groups truly differ on some measured outcome, and you happen to know the population standard deviations for both groups, you can use a two-sample -test. This situation is uncommon in practice (you rarely know ), but it builds the foundation for the more common -test procedures you'll see next.
The core question: Is the difference we observe between two sample means large enough to conclude the populations themselves differ, or could it just be sampling variability?
Test Statistic for Two Population Means
The test statistic measures how far the observed difference between sample means falls from the hypothesized difference, scaled by the standard error:
Here's what each piece represents:
- and : the sample means from group 1 and group 2
- : the hypothesized difference between population means under the null hypothesis (usually 0, meaning "no difference")
- and : the known population standard deviations
- and : the sample sizes for each group
Because the population standard deviations are known, this test statistic follows a standard normal distribution (the -distribution) when the null hypothesis is true.
How to use the result: Compare your calculated -value to the critical value set by your significance level. For a two-tailed test at , the critical value is . If , you reject the null hypothesis. You can also compare the -value to directly.
Quick example: Suppose you're comparing average test scores between two schools. School A has , , . School B has , , . Under :
-
Calculate the numerator:
-
Calculate the denominator:
-
Compute :
-
Since , you fail to reject at

Sampling Distribution of Mean Differences
The sampling distribution of describes all possible differences between sample means you could get if you repeated the sampling process over and over. Three key properties:
- Center: The mean of this distribution equals the true difference between population means:
- Spread: The standard deviation of this distribution (the standard error) is:
- Shape: The distribution is approximately normal when both sample sizes are large (typically and ), or when both populations are themselves normally distributed
This all assumes the two samples are independent, meaning the observations in one sample don't influence or relate to the observations in the other.

Standard Error of the Difference
The standard error tells you how much the difference tends to vary from sample to sample:
This is the denominator of the -test statistic, and it plays a critical role. A smaller standard error means your estimate of the difference is more precise, which makes it easier to detect a real difference if one exists. Two things shrink the standard error:
- Larger sample sizes ( and in the denominator)
- Smaller population standard deviations ( and in the numerator)
This is why researchers try to collect large samples: it tightens the standard error and gives the test more ability to pick up on real effects.
Statistical Considerations
- Pooled standard deviation: When you have reason to believe both populations share the same variance (), you can pool them into a single estimate. In the known- case this simplifies the formula, but it requires that equal-variance assumption to hold.
- Statistical power: This is the probability of correctly rejecting a false null hypothesis (detecting a real difference when one exists). Power increases with larger sample sizes, larger true effect sizes, and higher significance levels. Low power means you might miss a real difference, so planning adequate sample sizes before collecting data matters.