Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Policy evaluation isn't just an academic exercise—it's how governments decide whether your tax dollars are actually solving problems or getting wasted. You're being tested on your ability to distinguish between different evaluation approaches and, more importantly, to explain when and why each method is appropriate. The AP exam loves asking you to recommend an evaluation strategy for a given scenario or to identify the strengths and limitations of different approaches.
These methods fall into distinct categories based on what question they're trying to answer: Is this policy working? How do we know it's the policy causing the change? What do people actually think about it? How is implementation going? Don't just memorize the names—know what type of evidence each method produces and when evaluators would choose one approach over another.
These approaches attempt to establish causal relationships—proving that a policy actually caused an observed outcome rather than just being correlated with it. The key mechanism is comparison: what happened to people affected by the policy versus what would have happened without it.
Compare: RCTs vs. Regression Analysis—both seek causal evidence, but RCTs create comparison through randomization while regression creates it statistically. If an FRQ asks about evaluating a new job training program, RCTs give cleaner results but regression works when randomization isn't feasible.
These methods ask the fundamental question: Did the policy achieve what it was supposed to achieve? They focus on measuring results rather than understanding mechanisms.
Compare: CBA vs. Impact Assessment—CBA quantifies everything in dollars for direct comparison, while impact assessment captures qualitative and distributional effects that resist monetization. Use CBA when comparing budget alternatives; use impact assessment when equity concerns matter.
Sometimes policies fail not because they're bad ideas but because they're poorly implemented. These methods examine how programs operate in practice.
Compare: Process Evaluation vs. Program Evaluation—process evaluation zooms in on implementation fidelity, while program evaluation takes a broader view encompassing design and outcomes. Think of process evaluation as one component within a comprehensive program evaluation.
Policies don't exist in a vacuum—they affect real people whose perspectives matter both ethically and practically. These methods capture human experiences and viewpoints.
Compare: Surveys vs. Case Studies—surveys provide breadth (many respondents, standardized questions), while case studies provide depth (few cases, rich detail). An FRQ might ask you to explain when each approach is more appropriate—surveys for measuring prevalence, case studies for understanding mechanisms.
| Evaluation Question | Best Methods |
|---|---|
| Did the policy cause the outcome? | RCTs, Regression Analysis |
| Is the policy worth the investment? | Cost-Benefit Analysis |
| What effects did the policy produce? | Impact Assessment, Program Evaluation |
| Is the program running as intended? | Process Evaluation, Performance Measurement |
| What do affected people think? | Surveys, Stakeholder Analysis |
| How does implementation work in context? | Case Studies, Process Evaluation |
| Should we continue or expand the program? | Program Evaluation, CBA, Impact Assessment |
A city wants to know whether its new after-school program is actually reducing juvenile crime or whether crime was already declining. Which evaluation method would provide the strongest causal evidence, and why might it be difficult to implement?
Compare and contrast Cost-Benefit Analysis and Impact Assessment. In what situation would a policymaker choose one over the other?
Which two methods would you combine to understand both whether a policy is working and why it's working (or failing) in different locations?
A state agency has limited budget and needs to quickly assess public support for a proposed policy change. Which method is most appropriate, and what validity concerns should they address?
If an FRQ describes a program that's achieving good outcomes in some districts but poor outcomes in others, which evaluation approach would best explain this variation? What would it examine?