Confidence intervals are the bridge between your sample data and the truth about an entire population—and that's exactly what you're being tested on. Every time you calculate a CI, you're quantifying uncertainty, which is the heart of statistical inference. The AP exam loves testing whether you understand why intervals widen or narrow, when to use z versus t distributions, and how to interpret what a 95% confidence level actually means (hint: it's not "95% chance the parameter is in this interval").
The concepts here connect directly to hypothesis testing, sampling distributions, and the Central Limit Theorem. You'll need to recognize when conditions are met, choose the right formula, and—critically—interpret results in context. Don't just memorize formulas; know what each component controls and how changing sample size, confidence level, or variability affects your interval. Master the mechanics behind these calculations, and FRQs become much more manageable.
Foundational Concepts: What CIs Actually Measure
Before diving into formulas, you need to understand what confidence intervals represent. A CI captures the range of plausible values for a population parameter based on sample data, with a specified level of confidence.
Definition and Purpose
A confidence interval estimates a range where the true population parameter likely falls—it's not a probability statement about any single interval
The width reflects uncertainty—wider intervals mean less precision, narrower intervals mean more confidence in your estimate
CIs quantify sampling variability, connecting your sample statistic to the unknown population parameter through probability
Confidence Level and Margin of Error
The confidence level (e.g., 95%) means that if you repeated the sampling process many times, about 95% of the resulting intervals would contain the true parameter
Margin of error is the "±" portion—calculated as critical value×standard error
Higher confidence = wider interval, because you need a larger range to be "more sure" you've captured the parameter
Compare: Confidence level vs. margin of error—both affect interval width, but confidence level is chosen by the researcher while margin of error results from the calculation. On FRQs, always state your confidence level and interpret the margin of error in context.
Single-Parameter Estimation: Means and Proportions
These are your bread-and-butter calculations. The choice of formula depends on what you're estimating and what information you have about the population.
CI for Population Mean (Known σ)
Formula: xˉ±z∗⋅nσ where xˉ is the sample mean, σ is the known population standard deviation, and n is sample size
Use z-scores for critical values (e.g., z∗=1.96 for 95% confidence)—this assumes normality or n≥30 via the Central Limit Theorem
Rarely used in practice because knowing σ without knowing μ is uncommon—but it's a foundational concept for exams
CI for Population Mean (Unknown σ)
Formula: xˉ±t∗⋅ns where s is the sample standard deviation and t∗ comes from the t-distribution with df=n−1
The t-distribution has heavier tails than the normal distribution, accounting for the extra uncertainty when estimating σ from sample data
As n increases, the t-distribution approaches the normal distribution—this is why large samples can use either method
CI for Population Proportion
Formula: p^±z∗⋅np^(1−p^) where p^ is the sample proportion
Conditions required:np^≥10 and n(1−p^)≥10 for the normal approximation to be valid
The standard error np^(1−p^) is maximized when p^=0.5, which is why conservative sample size calculations use this value
Compare: z-interval for means vs. t-interval for means—both estimate μ, but z requires known σ while t uses sample s. If an FRQ doesn't give you σ, you're using t. Always state your degrees of freedom.
Factors That Control Interval Width
Understanding these relationships is crucial for exam questions that ask "what happens if...?" Interval width is determined by three factors: confidence level, sample size, and variability.
Effect of Sample Size
Increasing n decreases interval width because n appears in the denominator of the standard error
The relationship is square root, so quadrupling your sample size only halves the margin of error
Larger samples provide more information, reducing the uncertainty in your estimate of the population parameter
Relationship Between Confidence Level and Width
Higher confidence levels require larger critical values (z∗ or t∗), which directly increases the margin of error
99% CI is wider than 95% CI for the same data—you're trading precision for confidence
Choosing a confidence level involves balancing the need for certainty against the desire for a useful (narrow) interval
Sample Size Determination
For means: n=(Ez∗⋅σ)2 where E is the desired margin of error
For proportions: n=E2(z∗)2⋅p^(1−p^)—use p^=0.5 if no prior estimate exists
Always round up when calculating required sample size—you can't survey 384.2 people
Compare: Doubling confidence level vs. quadrupling sample size—increasing confidence widens the interval, while increasing sample size narrows it. FRQs often ask you to identify which change achieves a specific goal.
Comparing Two Populations
When you need to compare groups rather than estimate a single parameter, the formulas combine information from both samples. The standard error for a difference involves adding variances, not standard deviations.
Difference Between Two Means
Formula: (xˉ1−xˉ2)±t∗⋅n1s12+n2s22 for independent samples
Degrees of freedom can be calculated using Welch's approximation or conservatively as min(n1−1,n2−1)
If the interval contains zero, you cannot conclude a significant difference between the population means at that confidence level
Conditions: Both samples must independently satisfy np^≥10 and n(1−p^)≥10
Interpretation: If zero is not in the interval, evidence suggests a real difference between population proportions
Compare: Two-sample mean CI vs. two-sample proportion CI—both estimate differences, but means use the t-distribution while proportions use z. The structure of adding variances under the square root is identical.
Special Cases and Advanced Methods
These topics extend the basic framework to handle specific situations. Knowing when to apply each method demonstrates deeper statistical understanding.
One-Sided vs. Two-Sided Intervals
Two-sided CIs estimate parameters in both directions (e.g., μ could be higher or lower than xˉ)
One-sided CIs bound the parameter in only one direction—used when you only care about exceeding or falling below a threshold
Critical values differ: a one-sided 95% CI uses z∗=1.645, while two-sided uses z∗=1.96
Use of t-Distribution for Small Samples
Required when n<30 and population standard deviation is unknown—the CLT doesn't guarantee normality
Heavier tails compensate for the increased variability in estimating σ from small samples
Check the population distribution: if the underlying data is heavily skewed, even t-intervals may be unreliable for small n
Prediction Intervals
Formula: xˉ±t∗⋅s1+n1 estimates where a single new observation will fall
Wider than confidence intervals because they account for both sampling variability and individual observation variability
The "1 +" term captures the additional uncertainty of predicting one value rather than estimating a mean
Compare: Confidence interval vs. prediction interval—CIs estimate population parameters, PIs predict individual observations. PIs are always wider because individual values vary more than means.
Assumptions and Validity Conditions
No confidence interval is valid unless conditions are met. Checking assumptions isn't optional—it's required for full credit on FRQs.
Required Conditions
Random sampling: Data must be randomly selected from the population of interest
Independence: Observations must be independent; for sampling without replacement, check that n≤0.10N (10% condition)
Normality/sample size: For means, either the population is normal or n≥30; for proportions, check np^≥10 and n(1−p^)≥10
Bootstrap Method
Resampling with replacement creates thousands of simulated samples from your original data
No normality assumption required—the bootstrap builds an empirical sampling distribution
Particularly useful for complex statistics, small samples, or when traditional assumptions fail
Quick Reference Table
Concept
Best Examples
Known σ
z-interval for means, large population studies
Unknown σ
t-interval for means, most real-world applications
Proportion estimation
Single proportion CI, polling, survey analysis
Two-sample comparison
Difference of means, difference of proportions
Sample size planning
n=(z∗σ/E)2 for means, n=z2p^(1−p^)/E2 for proportions
Small samples
t-distribution, increased degrees of freedom sensitivity
Predicting individuals
Prediction intervals (wider than CIs)
Assumption-free methods
Bootstrap resampling
Self-Check Questions
When constructing a CI for a population mean, what two conditions determine whether you use a z-interval or a t-interval?
Compare and contrast: How does increasing the confidence level from 90% to 99% affect interval width, and how does quadrupling the sample size affect it? Which change narrows the interval?
A confidence interval for μ1−μ2 is (−2.3,4.7). What conclusion can you draw about whether the population means differ? Explain your reasoning.
Which two formulas both contain the term p^(1−p^), and why does this expression appear in both?
An FRQ asks you to construct a 95% CI for a proportion with n=40 and p^=0.08. Before calculating, what condition should you check, and is it satisfied? What are the implications?