Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
When randomized controlled trials aren't possible—due to ethical constraints, cost, or practical limitations—quasi-experimental designs become your primary toolkit for establishing causality. You're being tested on your ability to identify when each design is appropriate, what assumptions must hold, and how threats to validity differ across methods. These aren't just abstract techniques; they're the workhorses behind policy evaluation, program assessment, and empirical research in economics, public health, and social sciences.
Understanding these designs means recognizing that causal inference is fundamentally about ruling out alternative explanations. Each method addresses confounding in a different way—some exploit timing, others leverage cutoffs, and still others rely on external variation. Don't just memorize definitions—know what identifying assumption each design requires and what would cause it to fail.
These methods leverage the structure of when interventions occur to separate causal effects from pre-existing trends. The key insight: if you can observe the same units (or comparable units) before and after treatment, you can difference out confounding factors that don't change over time.
Compare: DiD vs. ITS—both exploit timing, but DiD requires a control group to difference out time trends while ITS relies on extrapolating pre-intervention trends. If you have a strong comparison group, DiD is preferred; if you have rich time-series data but no control, ITS is your fallback.
These methods identify causal effects by comparing units just above and below an arbitrary threshold. The logic: units near the cutoff are essentially randomly assigned to treatment, creating local randomization.
Compare: RDD vs. IV—both provide causal estimates without randomization, but RDD exploits a known assignment rule while IV exploits external variation. RDD gives you a clear visual test (plot the discontinuity); IV validity is harder to verify since the exclusion restriction cannot be directly tested.
When natural comparison groups don't exist, these methods create them statistically. The goal: balance observable characteristics between treated and control units to approximate what randomization would achieve.
Compare: PSM vs. Synthetic Control—both construct comparison groups, but PSM matches individual units while synthetic control creates a weighted composite. Use PSM when you have many treated units; use synthetic control when you're studying a single case (e.g., one state's policy change).
These approaches leverage real-world events that create quasi-random variation in treatment exposure. Nature or policy inadvertently runs the experiment for you.
Compare: Natural Experiments vs. Fixed Effects—natural experiments identify causal effects through external variation, while fixed effects control for stable confounders through within-unit comparisons. Natural experiments are about finding variation; fixed effects are about controlling for unobservables.
Not all causal inference is quantitative. These methods provide depth and context that statistical approaches may miss.
Compare: Comparative Case Studies vs. Quantitative Quasi-Experiments—case studies prioritize internal validity and mechanistic understanding within specific contexts, while quantitative methods prioritize estimating average effects across populations. Use case studies to understand why an effect occurs; use quantitative methods to estimate how large it is.
| Concept | Best Examples |
|---|---|
| Exploits timing/trends | DiD, Interrupted Time Series |
| Exploits cutoffs/thresholds | RDD, IV |
| Constructs comparison groups | PSM, Matching Methods, Synthetic Control |
| Controls for unobservables | Fixed Effects, IV, DiD |
| Single-unit or small-N studies | Synthetic Control, Comparative Case Studies |
| Requires parallel trends | DiD |
| Requires valid instrument | IV |
| Selection on observables only | PSM, Matching Methods |
Both DiD and ITS exploit timing to identify causal effects. What is the key difference in their data requirements, and when would you choose one over the other?
A researcher wants to estimate the effect of a scholarship program that's awarded to students scoring above 80 on an entrance exam. Which design is most appropriate, and what threat to validity should they test for?
Compare propensity score matching and instrumental variables: which assumption is stronger, and why might a researcher prefer IV despite its stricter requirements?
You're asked to evaluate a smoking ban implemented in one state. You have data on multiple states over 10 years. Which two methods could you use, and what are the tradeoffs between them?
FRQ-style: A policy analyst claims that fixed effects models "solve" the problem of omitted variable bias. Explain why this claim is only partially correct, identifying what types of confounders fixed effects can and cannot address.