Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Experimental design is the backbone of Unit 3 and shows up repeatedly throughout AP Statistics, from understanding how data should be collected to interpreting results in inference problems. When you see an FRQ asking whether a study can establish causation or merely association, you're being tested on these principles. The concepts here, randomization, control, blocking, and replication, aren't just vocabulary words. They're the tools that separate valid experiments from flawed ones.
Every design choice exists to solve a specific problem. Randomization eliminates confounding. Blocking reduces variability. Control groups provide a baseline. Blinding prevents bias. Don't just memorize these terms. Understand what problem each principle solves, because that's exactly what the exam will ask you to explain.
The fundamental goal of an experiment is to establish a cause-and-effect relationship. Three principles work together to make that possible, and without all three, your experiment cannot support causal conclusions.
Random assignment means each experimental unit has an equal chance of receiving any treatment, which creates roughly equivalent groups before the experiment even begins.
Having multiple experimental units per treatment lets you estimate the natural variability in responses and distinguish real effects from random noise.
A control group provides a baseline for comparison. Without knowing what happens in the absence of treatment, you can't measure the treatment's effect.
Compare: Randomization vs. Replication โ both are essential for valid experiments, but they solve different problems. Randomization creates comparable groups (controls confounding), while replication provides enough data to detect effects (controls variability). If an FRQ asks why an experiment can establish causation, mention both.
Even with perfect randomization, human psychology can introduce systematic errors. These techniques address bias that comes from participants' and researchers' expectations.
The placebo effect is a psychological response to perceived treatment. Participants may improve simply because they believe they're receiving help, not because the treatment actually works.
Compare: Single-blind vs. Double-blind โ single-blind controls participant bias only, while double-blind also prevents researchers from unconsciously treating groups differently or interpreting results favorably. Double-blind is the gold standard, but single-blind may be necessary when researchers must know treatments (e.g., surgical procedures).
When experimental units differ in ways that affect the response, blocking groups similar units together before randomization. This reduces noise in your data. Think of it as sorting before shuffling.
Here's how to set up a randomized block design:
This design is more precise than a completely randomized design when the blocking variable is strongly related to the response, because it removes that source of variability from the comparison.
Matched-pairs is a special case of blocking with only two treatments.
Compare: Randomized Block Design vs. Matched-Pairs โ both use blocking, but matched-pairs is specifically for two-treatment comparisons and often uses subjects as their own controls. On FRQs, identify matched-pairs when the same subjects receive both treatments or when subjects are explicitly paired before assignment.
Different research questions call for different experimental frameworks. The choice depends on how many factors you're studying, what resources you have, and the nature of your experimental units.
The simplest structure: all units are assigned to treatments purely by chance, with no blocking or matching.
A factorial design studies multiple factors simultaneously. For example, a factorial examines two factors, each at two levels, creating four treatment combinations total.
In a crossover design, each participant receives all treatments in sequence, serving as their own control. This dramatically reduces between-subject variability.
Compare: Completely Randomized vs. Randomized Block Design โ CRD is simpler but ignores known sources of variability, while RBD accounts for them through blocking. Choose CRD when units are homogeneous; choose RBD when you can identify a variable that affects the response. FRQs often ask you to justify why blocking improves an experiment.
Understanding what can go wrong helps you design better experiments and critique flawed studies. A common FRQ task is identifying problems in a described study.
A confounding variable is associated with both the treatment and the response, making it impossible to isolate the treatment's true effect. It provides an alternative explanation for the results.
Larger samples increase power, which is the ability to detect a real treatment effect when one exists.
Compare: Confounding vs. Bias โ both threaten validity, but they're different problems. Confounding is about alternative explanations (a third variable is related to both the treatment and the response). Bias is about systematic errors in measurement or selection. Randomization primarily addresses confounding; blinding primarily addresses bias.
| Concept | Best Examples |
|---|---|
| Establishing causation | Randomization, Control groups, Replication |
| Reducing psychological bias | Double-blind, Single-blind, Placebo control |
| Controlling known variability | Blocking, Randomized block design, Matched-pairs |
| Simple experimental structures | Completely randomized design |
| Complex factor analysis | Factorial design, Crossover design |
| Threats to validity | Confounding variables, Selection bias, Response bias |
| Precision and power | Replication, Sample size determination |
A researcher wants to test whether a new fertilizer increases tomato yield. She has 30 plants of varying ages and sizes. Should she use a completely randomized design or a randomized block design? Explain what blocking variable she might use and why it would improve the experiment.
What do randomization and blinding have in common, and how do they differ? Which one allows an experiment to establish causation, and which one prevents psychological bias?
An experiment compares two pain medications by giving each participant both drugs (one per week) in random order. What design is this, and why is a washout period necessary between treatments?
A study finds that coffee drinkers have lower rates of heart disease. A journalist claims coffee prevents heart disease. Explain why this conclusion is flawed and identify at least one potential confounding variable.
Compare matched-pairs design and randomized block design. When would you choose matched-pairs over blocking with multiple treatments? How does the analysis of matched-pairs data differ from analyzing two independent samples?