upgrade
upgrade

🫘Intro to Public Policy

Important Policy Evaluation Methods

Study smarter with Fiveable

Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.

Get Started

Why This Matters

Policy evaluation isn't just an academic exercise—it's how governments decide whether your tax dollars are actually solving problems or getting wasted. You're being tested on your ability to distinguish between different evaluation approaches and, more importantly, to explain when and why each method is appropriate. The AP exam loves asking you to recommend an evaluation strategy for a given scenario or to identify the strengths and limitations of different approaches.

These methods fall into distinct categories based on what question they're trying to answer: Is this policy working? How do we know it's the policy causing the change? What do people actually think about it? How is implementation going? Don't just memorize the names—know what type of evidence each method produces and when evaluators would choose one approach over another.


Experimental and Quasi-Experimental Methods

These approaches attempt to establish causal relationships—proving that a policy actually caused an observed outcome rather than just being correlated with it. The key mechanism is comparison: what happened to people affected by the policy versus what would have happened without it.

Randomized Controlled Trials (RCTs)

  • Gold standard for causal inference—randomly assigns participants to treatment and control groups, eliminating selection bias that plagues other methods
  • Produces the strongest evidence for policy effectiveness because randomization ensures the only systematic difference between groups is the intervention itself
  • High cost and ethical constraints limit applicability; you can't randomly deny people essential services, and implementation requires significant resources and time

Regression Analysis

  • Statistical technique isolating variable relationships—allows evaluators to control for confounding factors that might otherwise distort results
  • Enables before-and-after comparisons by analyzing pre- and post-implementation data while accounting for other changes occurring simultaneously
  • Requires quality data and careful specification—results are only as good as the variables included and the assumptions made about their relationships

Compare: RCTs vs. Regression Analysis—both seek causal evidence, but RCTs create comparison through randomization while regression creates it statistically. If an FRQ asks about evaluating a new job training program, RCTs give cleaner results but regression works when randomization isn't feasible.


Outcome-Focused Evaluation

These methods ask the fundamental question: Did the policy achieve what it was supposed to achieve? They focus on measuring results rather than understanding mechanisms.

Cost-Benefit Analysis (CBA)

  • Compares total costs to total benefits in monetary terms—forces explicit consideration of trade-offs and opportunity costs
  • Enables comparison across different policy options by converting diverse outcomes (health improvements, time saved, environmental quality) to a common metric
  • Subjectivity in valuing intangibles creates controversy; how do you assign dollar values to a life saved or a species preserved?

Impact Assessment

  • Evaluates broader effects on target populations and environment—goes beyond intended outcomes to capture ripple effects throughout the system
  • Captures unintended consequences both positive and negative, providing a comprehensive picture of what the policy actually changed
  • Informs policy redesign by identifying where adjustments could maximize benefits or minimize harms

Performance Measurement

  • Tracks key performance indicators (KPIs) over time—establishes benchmarks and monitors whether programs are hitting their targets
  • Enables accountability and transparency by giving stakeholders concrete metrics to evaluate program success
  • Focuses on outputs and outcomes rather than causes; tells you what is happening but not necessarily why

Compare: CBA vs. Impact Assessment—CBA quantifies everything in dollars for direct comparison, while impact assessment captures qualitative and distributional effects that resist monetization. Use CBA when comparing budget alternatives; use impact assessment when equity concerns matter.


Process and Implementation Evaluation

Sometimes policies fail not because they're bad ideas but because they're poorly implemented. These methods examine how programs operate in practice.

Process Evaluation

  • Examines whether programs operate as designed—identifies gaps between policy on paper and policy in practice
  • Reveals barriers and facilitators to effective implementation, explaining why some sites succeed while others struggle
  • Essential for program improvement because you can't fix what you don't understand; tells you where the breakdown is occurring

Program Evaluation

  • Comprehensive assessment of design, implementation, and outcomes—combines multiple methods to build a complete picture
  • Uses mixed methods approach integrating both quantitative data (numbers served, costs incurred) and qualitative data (participant experiences, staff perspectives)
  • Generates evidence-based recommendations for improving current programs and designing future initiatives

Compare: Process Evaluation vs. Program Evaluation—process evaluation zooms in on implementation fidelity, while program evaluation takes a broader view encompassing design and outcomes. Think of process evaluation as one component within a comprehensive program evaluation.


Stakeholder and Perception Methods

Policies don't exist in a vacuum—they affect real people whose perspectives matter both ethically and practically. These methods capture human experiences and viewpoints.

Surveys and Questionnaires

  • Collect standardized data from large populations—gauge opinions, behaviors, and experiences related to policy impacts
  • Inform policy design through public input by revealing what constituents actually want and need
  • Validity depends on careful design—poorly worded questions, biased samples, or low response rates can produce misleading results

Stakeholder Analysis

  • Maps interests, influence, and needs of individuals and groups affected by a policy
  • Identifies potential allies and opponents helping policymakers anticipate resistance and build coalitions
  • Ensures diverse perspectives inform decisions—policies designed without stakeholder input often face implementation challenges

Case Studies

  • Deep examination of specific implementation instances—explores how policies play out in real-world contexts with all their complexity
  • Generates rich qualitative understanding of mechanisms, context, and nuance that quantitative methods miss
  • Complements statistical approaches by explaining the "why" behind the numbers and identifying factors that surveys might overlook

Compare: Surveys vs. Case Studies—surveys provide breadth (many respondents, standardized questions), while case studies provide depth (few cases, rich detail). An FRQ might ask you to explain when each approach is more appropriate—surveys for measuring prevalence, case studies for understanding mechanisms.


Quick Reference Table

Evaluation QuestionBest Methods
Did the policy cause the outcome?RCTs, Regression Analysis
Is the policy worth the investment?Cost-Benefit Analysis
What effects did the policy produce?Impact Assessment, Program Evaluation
Is the program running as intended?Process Evaluation, Performance Measurement
What do affected people think?Surveys, Stakeholder Analysis
How does implementation work in context?Case Studies, Process Evaluation
Should we continue or expand the program?Program Evaluation, CBA, Impact Assessment

Self-Check Questions

  1. A city wants to know whether its new after-school program is actually reducing juvenile crime or whether crime was already declining. Which evaluation method would provide the strongest causal evidence, and why might it be difficult to implement?

  2. Compare and contrast Cost-Benefit Analysis and Impact Assessment. In what situation would a policymaker choose one over the other?

  3. Which two methods would you combine to understand both whether a policy is working and why it's working (or failing) in different locations?

  4. A state agency has limited budget and needs to quickly assess public support for a proposed policy change. Which method is most appropriate, and what validity concerns should they address?

  5. If an FRQ describes a program that's achieving good outcomes in some districts but poor outcomes in others, which evaluation approach would best explain this variation? What would it examine?