Study smarter with Fiveable
Get study guides, practice questions, and cheatsheets for all your subjects. Join 500,000+ students with a 96% pass rate.
Teacher evaluation systems sit at the intersection of several major policy debates you'll encounter throughout this course: accountability versus autonomy, quantitative versus qualitative assessment, and the tension between standardization and professional judgment. Understanding how these systems work—and why they generate so much controversy—helps you analyze broader questions about how we measure educational quality and what incentives shape teacher behavior.
You're being tested on your ability to evaluate policy mechanisms, not just describe them. When you see teacher evaluation on an exam, you need to understand why policymakers chose certain approaches, what trade-offs each method involves, and how different stakeholders experience these systems. Don't just memorize the names of evaluation frameworks—know what problem each one tries to solve and what limitations it creates.
These methods attempt to isolate teacher effectiveness using numerical data, appealing to policymakers who want objective, comparable metrics. The core assumption is that student performance data can reveal teacher quality when properly analyzed.
Compare: Value-Added Models vs. Student Growth Measures—both use student data to assess teachers, but VAM attempts complex statistical controls while growth measures focus more simply on progress over time. If an FRQ asks about data-driven evaluation, distinguish between these approaches and their different assumptions about what "counts" as evidence.
These methods prioritize professional judgment and direct evidence of teaching practice. The underlying principle is that effective teaching involves complex behaviors that can only be assessed through trained observation and documentation.
Compare: Classroom Observations vs. Portfolio Assessments—both rely on qualitative evidence, but observations capture real-time practice while portfolios document curated, reflective work. Observations risk the "dog and pony show" problem; portfolios risk selective presentation.
These approaches recognize that different participants in the educational process have unique insights into teacher effectiveness. The theory is that triangulating perspectives produces a more complete picture than any single viewpoint.
Compare: Student Surveys vs. Peer Evaluations—both gather stakeholder perspectives, but students report on their direct experience as learners while peers evaluate professional practice. Students see things administrators miss; peers understand instructional nuances outsiders can't assess.
These comprehensive models provide shared language and criteria for evaluation, attempting to codify effective teaching into observable, measurable components. They address the fundamental policy question: what exactly should we be looking for when we evaluate teachers?
Compare: Danielson Framework vs. Marzano Model—both provide structured rubrics for evaluation, but Danielson emphasizes broader professional practice and reflection while Marzano focuses more specifically on instructional strategies and their research base. Know which framework your state or district uses—it shapes how "good teaching" gets defined locally.
These concepts address how evaluation components combine into coherent systems and connect to broader policy goals. The design question is how to balance competing values: accuracy, fairness, feasibility, and improvement.
Compare: Multiple Measures Approach vs. Performance-Based Compensation—multiple measures is an evaluation design philosophy, while performance-based pay is a policy application of evaluation results. You can use multiple measures without tying them to compensation, but compensation systems almost always require multiple measures to be defensible.
| Concept | Best Examples |
|---|---|
| Data-driven evaluation | Value-Added Models, Student Growth Measures |
| Observation-based assessment | Classroom Observations, Danielson Framework, Marzano Model |
| Stakeholder voice | Student Surveys, Peer Evaluations |
| Documentation of practice | Portfolio Assessments |
| Comprehensive frameworks | Danielson Framework, Marzano Model |
| System design principles | Multiple Measures Approach |
| Incentive structures | Performance-Based Compensation Systems |
| Quantitative methods | VAM, Student Growth Measures |
| Qualitative methods | Observations, Portfolios, Peer Evaluations |
Which two evaluation methods both rely on student data but differ in their statistical complexity and assumptions about isolating teacher effects?
A district wants to reduce subjectivity in teacher evaluation while still capturing the complexity of classroom practice. Which combination of methods would best address both concerns, and what trade-offs would remain?
Compare and contrast the Danielson Framework and Marzano Model: what do they share as structured observation tools, and how do they differ in emphasis and design philosophy?
If an FRQ asks you to evaluate the equity implications of teacher evaluation systems, which methods would you identify as most problematic for teachers in high-poverty schools, and why?
A teachers' union argues that evaluation should support professional growth rather than determine employment consequences. Which evaluation methods align best with this developmental purpose, and which seem designed primarily for accountability?