When It Comes To How We Use Evidence, Is Education Reform The New Welfare Reform?

Posted by on June 2, 2011

** Also posted here on “Valerie Strauss’ Answer Sheet” in the Washington Post

In the mid-1990s, after a long and contentious debate, the U.S. Congress passed the Personal Responsibility and Work Opportunity Reconciliation Act of 1996, which President Clinton signed into law. It is usually called the “Welfare Reform Act,” as it effectively ended the Aid to Families with Dependent Children (AFDC) program (which is what most people mean when they say “welfare,” even though it was [and its successor is] only a tiny part of our welfare state). Established during the New Deal, AFDC was mostly designed to give assistance to needy young children (it was later expanded to include support for their parents/caretakers as well).

In place of AFDC was a new program – Temporary Assistance for Needy Families (TANF). TANF gave block grants to states, which were directed to design their own “welfare” programs. Although the states were given considerable leeway, their new programs were to have two basic features: first, for welfare recipients to receive benefits, they had to be working; and second, there was to be a time limit on benefits, usually 3-5 years over a lifetime, after which individuals were no longer eligible for cash assistance (states could exempt a proportion of their caseload from these requirements). The general idea was that time limits and work requirements would “break the cycle of poverty”; recipients would be motivated (read: forced) to work, and in doing so, would acquire the experience and confidence necessary for a bootstrap-esque transformation.

There are several similarities between the bipartisan welfare reform movement of the 1990s and the general thrust of the education reform movement happening today. For example, there is the reliance on market-based mechanisms to “cure” longstanding problems, and the unusually strong liberal-conservative alliance of the proponents. Nevertheless, while calling education reform “the new welfare reform” might be a good soundbyte, it would also take the analogy way too far.

My intention here is not to draw a direct parallel between the two movements in terms of how they approach their respective problems (poverty/unemployment and student achievement), but rather in how we evaluate their success in doing so. In other words, I am concerned that the manner in which we assess the success or failure of education reform in our public debate will proceed using the same flawed and misguided methods that were used by many for welfare reform.

I often hear policymakers and pundits of both parties assert, without a trace of doubt, that welfare reform was a success. For years after the law’s enactment, their evidence seemed to consist of one thing: The number of people on welfare rolls was declining. The interpretation of these data was that the law was helping people move from “welfare to work.” As the years went on, supporters of welfare reform also pointed to declining poverty rates as evidence of success.

Both inferences are, by themselves, unsupported at best. Sure, it’s conceivable that declining welfare rolls might signal that most recipients are moving off public assistance and into gainful employment. But, given the fact that the law was specifically designed to cut the welfare rolls (by imposing work requirements and time limits on benefits), an equally plausible interpretation is that people were being forced off welfare into deeper poverty. Similarly, it is entirely possible that the coincidence of shrinking welfare rolls and declining poverty rates reflects a causal connection, but – given the fact that we were undergoing an economic boom at the time – it’s more likely that the decline in poverty was a function of overall economic conditions and not of a relatively small program affecting poor single mothers.

To truly evaluate the effects of welfare reform, it is necessary to go far beyond improper causal inferences drawn from deficient data. When you do, you get a somewhat different picture. In fact, it has been shown that welfare reform, while no doubt successful in many cases, actually acted to weaken the safety net for the poorest Americans, while a decline in the overall poverty rate masked an increase in the nation’s rate of “deep poverty.”  There were unintended consequences, such as decreased college attendance among poor mothers who had to work for their benefits. And, finally, much of the alleged success was due more to the strong economy of the mid- to late-1990s than to TANF itself, as is evident in lower success rates among recent welfare “leavers” compared with the late 1990s, as well as the failures of many states’ TANF programs to respond adequately to the current recession.

We can debate whether the welfare reform law was successful – and there are many studies on this topic, covering a wide variety of outcomes – but the point is that the “conventional wisdom” that it was a smashing success, fueled by mostly-positive media coverage, was based on crude interpretations of inadequate evidence.

I’m concerned that we’re headed down this same road in education policy. For example, let’s say, hypothetically, that Florida, which just enacted huge changes to teacher personnel policy, makes a significant gain on its 2013 NAEP results. You can already hear the exultant cries: The reforms worked! Other states must follow suit!

First, as is the case with declining welfare rolls, an increase in cross-sectional test scores does not necessarily signal actual progress (it could be, for instance, the result of cohort differences), and even if the gains are “real,” they may not always signal greater readiness for college and career. Moreover, just like poverty rates and welfare reform, it would be hard, absent rigorous evaluation, to say whether it is Florida’s teacher-focused reforms themselves that were responsible, or general conditions such as higher standards and more effective administration. Finally, in much the same way that poverty rates can increase while “deep poverty” rates can go up, overall test score averages mask underlying distributional differences – by income, language ability, race, achievement level, and a dozen other things – that are rarely accorded any attention in our public discourse.

But I don’t need hypothetical examples. Our welfare reform-style evidentiary standards are all around us. Rising test scores are reflexively attributed to the radical changes in New Orleans public schools after Hurricane Katrina, ignoring the possibility that the largest displacement since the Civil War had changed the composition of the city’s public school students. Author Richard Whitmire has spent the past year or so traveling the country, offering simplistic misinterpretations of testing data as “evidence” of Michelle Rhee’s effectiveness (as has Rhee herself). Our entire charter school debate is a never-ending battle between proponents and opponents determined to show (often improperly) that charters do or do not boost test scores – with almost nobody asking the important question: Why? In fact, to whatever degree NCLB was spurred along by the “Texas miracle” (for example, “miracle” graduation rates in Houston were actually due to unrecorded dropouts), our primary federal education policy was enabled by a misreading of flawed data.

This needs to stop, on both “sides” of the debate. It took over a decade for pundits and policymakers to start asking serious questions about the true effects of welfare reform, and, even today, very few people beyond the research community have seen or examined the evidence (which is, by the way, still growing – you cannot evaluate a policy intervention solely in the short-term). Instead, declining rolls and poverty rates were used in much the same way as many use test scores: Incorrectly.

We can and do disagree about the merits of the new policies now being enacted throughout the nation. But there should be no disagreement as to the necessity of monitoring their impacts rigorously and continuously. Let’s point out when data are interpreted improperly, even when it doesn’t support our own point of view. Let’s look at a range of outcomes, not just test scores, and certainly not just proficiency rates. And let’s not make causal arguments based on any analysis that is not specifically designed to identify a causal impact.

Some policies work, others don’t, and there’s nothing wrong with trial and error, so long as you know which is which.

- Matt Di Carlo


5 Comments posted so far

  • This is smart and spot on and I highly recommend reading robin Rogers dillon book the welfare experiments, then consider the RCT movement…. Ive blogged on this at eduoptimists — see archives for higher Ed copouts

    Comment by Sara Goldrick Rab
    June 2, 2011 at 10:01 AM
  • Overall a good post, but it would be useful if you’d occasionally call out your own side for egregious offenses against the principles of social science. For example, Diane Ravitch has often used NAEP trends that don’t control for anything to make causal claims about the effectiveness of vouchers in Milwaukee, the effectiveness of unions in different states, and the effectiveness of charter schools. Perhaps someone on her own side would be able to keep her honest.

    Comment by Stuart Buck
    June 3, 2011 at 10:28 AM
  • Mr. Di Carlo-

    Florida was administering the 2011 NAEP round about the same time that the teacher bill was passing the Florida legislature. I can promise you therefore that you will not be hearing any claims from me that either the teacher bill or any other 2011 bills had anything to do with whatever happens with the 2011 NAEP.

    We can both monitor state tests over the next few years, and see what happens in the 2013 and 2015 NAEP exams. Policies take hold over time, and you are correct that one must be mindful of multiple other changes going on at the same time.

    In the meantime, there is more than enough high-quality statistical evidence, much of it from left of center sources such as Brookings, to convince a reasonable person that we ought to be revising the way we recruit, compensate and retain teachers. Florida will not be the only case we have to examine going forward.

    I am hopeful that we will see a variety of different approaches tried, and learn a great deal.

    Comment by Matthew Ladner
    June 3, 2011 at 6:35 PM
  • Thanks for the comment, Matthew. You’re of course correct about the 2011 NAEP, which was a careless mistake. I meant 2013, and I am correcting the text to reflect this change. I apologize for the error, and thanks for pointing it out.

    As for your main points, I agree on all counts.

    Thanks again,
    MD

    Comment by Matthew Di Carlo
    June 4, 2011 at 3:55 PM
  • [...] much less “powerful evidence” – of IMPACT’s success is a picture-perfect example of the flawed evidentiary standards that too often drive our education debate. The unfortunate reality is that we have virtually no [...]

    July 27, 2011 at 8:46 AM

Sorry, the comment form is closed at this time.

Disclaimer

This web site and the information contained herein are provided as a service to those who are interested in the work of the Albert Shanker Institute (ASI). ASI makes no warranties, either express or implied, concerning the information contained on or linked from shankerblog.org. The visitor uses the information provided herein at his/her own risk. ASI, its officers, board members, agents, and employees specifically disclaim any and all liability from damages which may result from the utilization of the information provided herein. The content in the shankerblog.org may not necessarily reflect the views or official policy positions of ASI or any related entity or organization.

Banner image adapted from 1975 photograph by Jennie Shanker, daughter of Albert Shanker.