Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Regression Discontinuity Design sits at the heart of modern causal inference because it solves a fundamental problem: how do we estimate causal effects when we can't randomly assign treatment? RDD exploits arbitrary cutoffs—test score thresholds, age limits, income eligibility lines—to create conditions that approximate a randomized experiment. You're being tested on your ability to recognize when RDD is appropriate, understand what makes it credible, and identify threats to its validity.
The concepts here connect directly to broader themes in causal inference: identification strategies, local versus average treatment effects, and the bias-variance tradeoff in estimation. Don't just memorize that "bandwidth matters" or "manipulation is bad"—understand why observations near a cutoff serve as valid counterfactuals and what assumptions must hold for that logic to work. These distinctions will drive your FRQ responses and help you critically evaluate research designs.
RDD works because individuals just above and below a cutoff are essentially randomly assigned to treatment—at least locally. The key insight is that if the cutoff is arbitrary, people on either side should be comparable in all ways except their treatment status.
Compare: Sharp RDD vs. Fuzzy RDD—both exploit the same cutoff logic, but sharp designs estimate treatment effects directly while fuzzy designs estimate effects only for those whose treatment status changed because of the threshold. If an FRQ describes imperfect compliance with a cutoff rule, you're dealing with fuzzy RDD.
The credibility of any RDD hinges on whether the discontinuity in treatment is the only thing changing at the cutoff. These assumptions determine whether your causal claims are defensible.
Compare: Manipulation testing vs. Covariate balance—both assess validity but target different threats. Density tests catch sorting into treatment; covariate checks catch discontinuities in baseline characteristics. Strong RDD papers report both.
How you estimate the treatment effect matters enormously. The core tension: using more data reduces variance but risks bias from observations far from the cutoff.
Compare: Wide bandwidth vs. Narrow bandwidth—wider bandwidths give you more statistical power but assume the relationship between the running variable and outcome is correctly specified far from the cutoff. Narrow bandwidths are more credible but noisier. Always report results across multiple bandwidths.
No single RDD estimate should be taken at face value. Credible research demonstrates that findings survive alternative specifications and acknowledges inherent limitations.
Compare: RDD limitations vs. RCT limitations—RCTs offer broader internal validity but face external validity concerns about artificial settings. RDD has strong local validity but explicitly cannot speak to effects away from the cutoff. Know which limitation matters more for a given research question.
Understanding where RDD fits in the causal inference toolkit helps you recognize when it's the right method—and when alternatives might be stronger.
Compare: RDD vs. Difference-in-Differences—RDD exploits a cross-sectional discontinuity in a running variable, while DiD exploits a temporal discontinuity (before/after treatment). RDD requires continuity assumptions; DiD requires parallel trends. Choose based on your data structure and which assumptions are more plausible.
| Concept | Best Examples |
|---|---|
| Design types | Sharp RDD, Fuzzy RDD |
| Core assumptions | Continuity, No manipulation, Running variable continuous |
| Validity tests | McCrary density test, Covariate balance checks, Placebo cutoffs |
| Estimation choices | Bandwidth selection, Local linear regression, Polynomial specifications |
| Robustness strategies | Sensitivity analysis, Multiple bandwidths, Alternative functional forms |
| Key limitations | Local effects only, External validity concerns, Single-cutoff dependence |
| Common applications | Education (test scores), Policy (eligibility thresholds), Health (age cutoffs) |
| Related methods | Instrumental variables (fuzzy RDD), RCTs (benchmark), Matching (alternative) |
What distinguishes sharp RDD from fuzzy RDD, and how does this distinction affect what parameter you're estimating?
A researcher finds that predetermined covariates show a discontinuity at the cutoff. What does this suggest about the validity of the RDD, and what assumption is likely violated?
Compare and contrast bandwidth selection in RDD with the bias-variance tradeoff: why might a researcher report results across multiple bandwidths rather than choosing a single "optimal" one?
If an FRQ asks you to evaluate a study using RDD to estimate the effect of a scholarship on college completion, what three validity checks would you look for in the research design?
Why does RDD estimate a Local Average Treatment Effect rather than an Average Treatment Effect, and what does this imply for generalizing findings to other populations?