Read the whole thing @Shanker BlogMuch of the criticism of value-added (VA) focuses on systematic bias, such as that stemming from non-random classroom assignment (also here). But the truth is that most of the imprecision of value-added estimates stems from random error. Months ago, I lamented the fact that most states and districts incorporating value-added estimates into their teacher evaluations were not making any effort to account for this error. Everyone knows that there is a great deal of imprecision in value-added ratings, but few policymakers seem to realize that there are relatively easy ways to mitigate the problem.
This is the height of foolishness. Policy is details. The manner in which one uses value-added estimates is just as important – perhaps even more so – than the properties of the models themselves. By ignoring error when incorporating these estimates into evaluation systems, policymakers virtually guarantee that most teachers will receive incorrect ratings. Let me explain.
Each teacher’s value-added estimate has an error margin (e.g., plus or minus X points). Just like a political poll, this error margin tells us the range within which that teacher’s “real” effect (which we cannot know for certain) falls. Unlike political polls, which rely on large random samples to get accurate estimates, VA error margins tend to be gigantic. One concrete example is from New York City, where the average margin of error was plus or minus 30 percentile points. This means that a New York City teacher with a rating at the 60th percentile might “actually” be anywhere between the 30th and 90th percentiles. We cannot even say with confidence whether this teacher is above or below average.
...
[Update! I forgot the following paragraph!!]
Now, here’s the problem: In virtually every new evaluation system that incorporates a value-added model, the teachers whose scores are not significantly different from the average are being treated as if they are. For example, some new systems sort teachers by their value-added scores, and place them into categories – e.g., the top 25 percent are “highly effective,” the next 25 percent are “effective,” the next 25 percent are “needs improvement,” and the bottom 25 percent are “ineffective.”
Showing posts with label assessment. Show all posts
Showing posts with label assessment. Show all posts
5/31/11
Value Added Measures: Fail (Shanker Blog): Updated
9/20/10
Report Cards Just Might Suck
Doctors now get report cards as a way to focus in on medical interventions to save money, not heal people. Sound familiar? Click on the Larry Cuban link for the whole thing.
The quarterly “report card” sits on my desk. Only 33% of my patients with diabetes have glycated hemoglobin levels that are at goal. Only 44% have cholesterol levels at goal. A measly 26% have blood pressure at goal. All my grades are well below my institution's targets.Danielle Ofri, M.D., Ph.D., via Larry Cuban
It's hard not to feel like a failure when the numbers are so abysmal. We've been getting these reports for more than 2 years now, and my numbers never budge. It's wholly dispiriting.
8/10/09
What's Wrong With Public Schools? Principals!
This memo, from my principal a year ago, shows just how the process works for diagnosing and getting services (assessment; the other kind) for a student who may be in need. I have redacted (like the government) the principal's name and initials, as well as some other identifying material.
Note how the principal makes sure to include the instruction to teachers: "It is not appropriate to say, "I think...""
Well, I do think, so that's going to be a bit difficult, no?
My hand-written question was merely rhetorical.
No, she never established an online anything.
Note how the principal makes sure to include the instruction to teachers: "It is not appropriate to say, "I think...""
Well, I do think, so that's going to be a bit difficult, no?
My hand-written question was merely rhetorical.
No, she never established an online anything.
Subscribe to:
Posts (Atom)