National Academies Press: OpenBook

Best Practices for State Assessment Systems, Part I: Summary of a Workshop (2010)

Chapter: 4 Political Experiences and Considerations

« Previous: 3 Innovative Assessment - Lessons from the Past and Present
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

4
Political Experiences and Considerations

As the recurring theme of tradeoffs suggests, political considerations are a very important aspect of almost all decisions about assessment and accountability, and they have played a critical role in the history of states’ efforts with innovative assessments. Veterans of three programs—the Maryland School Performance Assessment Program (MSPAP), the Kentucky Instructional Results Information System (KIRIS), and the Minnesota Comprehensive Assessment (MCA) in science—reflected on the political challenges of implementing these programs and the factors that have affected the outcomes for each.

MARYLAND

MSPAP was a logical next step for a state that had been pursuing education reform since the 1970s, Steve Ferrara explained. Maryland was among the first to use school- and system-level data for accountability, and the state also developed the essay-based Maryland Writing Test in the 1980s. That test, one of the first to use human scoring, set the stage for MSPAP. MSPAP was controversial at first, particularly after the first administration yielded pass rates as low as 50 percent for 9th graders. However, the test had a significant influence on writing instruction, and scores quickly rose.

The foundation for MSPAP was a Commission on School Performance established in 1987 by then governor William Schaeffer, which outlined ambitious goals for public education and recommended the establishment of content standards that would focus on higher order thinking skills. The commission also explicitly recommended that the state adopt an assessment that would not

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

rely only on multiple-choice items, and could provide annual report cards for schools.

The governor’s leadership was critical in marshalling the support of business and political leaders in the state, in Ferrara’s view. Because the Maryland governor appoints members of the state board of education, who, in turn, appoint the superintendent of public instruction, the result was “a team working together on education reform.” This team was responding to shifting expectations for education nationally, as well as a sense that state and district policy makers and the public were demanding assurances of the benefits of their investment in education. Ferrara recalled that the initial implementation went fairly smoothly, owing in part to concerted efforts by the state superintendent and others to communicate clearly with the districts about the goals for the program and how it would work and to solicit their input.

Most school districts were enthusiastic supporters, but several challenges complicated the implementation. The standards focused on broad themes and conceptual understanding, and it was not easy for test developers to design tasks that would target those domains in a way that was feasible and reliable. The way the domains were described in the standards led to complaints that the assessment did not do enough to assess students’ factual knowledge. The schedule was also exceedingly rapid, with just 11 months between initial planning and the first administration. The assessment burden was great—9 hours over 5 days for 3rd graders, for example. There were also major logistical challenges posed by the manipulatives needed for the large number of handson items.

The manipulatives not only presented logistical challenges, they also revealed a bigger challenge. For example, teachers who had not even been teaching science were asked to lead students through an assessment that included experiments. Teachers were also being asked more generally to change their instruction. The program involved teachers in every phase—task development, scoring, etc.—and Ferrara believes that this was one of the most important ingredients in its early success. As discussed above, there is evidence that teachers changed their practice in response to MSPAP (Koretz et al., 1996; Lane, 1999). Nevertheless, many people in the state began to oppose the program. Criticisms of the content and concern about the lack of individual student scores were the most prominent complaints, and the passage of the No Child Left Behind (NCLB) Act in 2002 made the latter concern urgent. The test was discontinued in that year.

For Ferrara, several key lessons can be learned from the history of MSPAP:

  • Involving stakeholders in every phase of the process was very valuable, both improving the quality of the program and building political acceptance.

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
  • It paid to be ambitious technically, but it is not necessary to do everything at once. For example, an assessment program could have a significant influence on instruction without being exclusively composed of open-ended items. If one makes a big investment in revolutionary science assessment, there will be fewer resources and less political capital for other content areas.

  • It was short-sighted to invest the bulk of funds and energy in test development at the expense of ongoing professional development.

KENTUCKY

The 1989 Kentucky Supreme Court decision that led to the development of KIRIS was intended to address both stark inequities in the state’s public education system and the state’s chronic performance at or near the bottom among the 50 states, Brian Gong explained. The resulting Kentucky Education Reform Act (KERA), passed in 1990 (which had broad bipartisan and public support), was one of the first state education reform bills and included innovative features, such as a substantial tax increase to fund reform; a restructuring of education governance; the allocation of 10 paid professional development days per year for teachers; and a revamped standards, assessment, and accountability system.

KERA established accountability goals for schools and students (proficiency within 20 years) and called for an assessment system that would be standards based, would rely primarily on performance-based items, and would be contextualized in the same way high-quality classroom instruction is. The result, KIRIS, met these specifications, but the developers faced many challenges. Most critically, the court decision had actually identified the outcomes to be measured by the assessment. These were the basis for the development of academic expectations and the core content for assessment, but the definitions of the constructs remained somewhat elusive. Educators complained that they were not sure what they were supposed to be doing in the classroom, Gong explained, and in the end it was the assessment that defined the content that was valued, the degree of mastery that was expected, and the way students would demonstrate that mastery. But developing tasks to assess the standards in valid ways was difficult, and the assessment alone could not provide sufficient information to influence instruction.

There were other challenges and adaptations. It was difficult to collect data for standard setting using the complex, multifaceted evidence of students’ skills and knowledge that KIRIS was designed to elicit. Equating the results from year to year was also difficult: most of the tasks were very memorable and could not be repeated. Alternate strategies—such as equating to administrations in other states, judgmental equating, and equating to multiple-choice items—had psychometric and practical disadvantages. KIRIS also initially struggled to main-

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

tain standards of accuracy and reliability in scoring the constructed-response items and portfolios, and adaptations and improvements were made in response to problems. Guidelines for the development of portfolios had to be strengthened in response to concerns about whether they truly reflected students’ work. With experience, test developers also gradually moved from holistic scoring of portfolios to analytic scoring in order to gain more usable information from the results. The state also faced logistical challenges, for example, with field testing and providing results in time for accountability requirements.

Nontechnical challenges emerged as well. The state was compelled to significantly reduce its assessment staff and to rely increasingly on consultants. School accountability quickly became unpopular, Gong explained, and many people began to complain that the aspirations were too high and to question the assertion that all students could learn at such high levels. Philosophical objections to the learning outcomes assessed by KIRIS emerged, with some arguing that many of them intruded on parents’ prerogatives and invaded students’ privacy. The so-called math and reading “wars”—over the relative emphasis that should be given to basic skills and fluency as opposed to broader cognitive objectives—fueled opposition to KIRIS. There were changes in the state’s political leadership that decreased support for the program, which did not survive an increasingly contentious political debate and was ended in 1998.

For Gong there are several key lessons:

  • Clear definitions of the constructs to be measured and the purposes and uses of the assessment are essential. No assessment can compensate for poorly defined learning targets.

  • The design of the assessment should be done in tandem with the design of the intended uses of the data, such as accountability, so that they can be both coherent and efficient.

  • The people who are proposing technical evaluations and those who will be the subject of them should work together in advance to consider both intended and unintended consequences, particularly in a politically charged context.

  • Anyone now considering innovative assessments for large-scale use should have a much clearer idea of how to meet technical and operational challenges than did the pioneering developers of KIRIS in the 1990s.

  • Current psychometric models, which support traditional forms of testing, are inconsistent with new views of both content and cognition and should be applied only sparingly to innovative assessments. The field should invest in the development of improved models and criteria (see, e.g., Shepard, 1993; Mislevy, 1998).

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

MINNESOTA

Minnesota’s Comprehensive Assessment Series II (MCA-II) was developed in response to NCLB, so the state was able to benefit from the experiences of states that had already explored innovative assessment. Assessments already in place in some subjects could be adapted, but the MCA-II in science, implemented in 2008, presented an opportunity to do something new. State assessment staff were also given unusual latitude to experiment, Dirk Mattson explained, because science had not been included in the NCLB accountability requirements.1

The result was a scenario-based assessment delivered on computers. Students are presented with realistic representations of classroom experiments, as well as phenomena that can be observed. Items—which may be multiple choice, short or long constructed-response, or figural (e.g., students interact with graphics in some way)—are embedded in the scenario. This structure provides students with an opportunity to engage in science at a higher cognitive level than would be possible with stand-alone items.

Mattson emphasized that the design of the science assessment grew out of the extensive involvement of teachers from the earliest stages. Teachers rejected the idea of an exclusively multiple-choice test, but a statewide performance assessment was also not considered because a previous effort had ended in political controversy. The obvious solution was a computer-delivered assessment, despite concerns about the availability of the necessary technology in schools. The developers had the luxury of a relatively generous schedule. Conceptual design began in 2005, and the first operational assessment was given in 2008. This allowed time for some pilot testing at the district level before field testing began in 2007.

A few complications have arisen. First, Minnesota statute requires regular revision of education standards, so the science standards were actually being revised before the prior ones had been assessed, but assessment revision was built into the process. Second, in 2009 the state legislature, facing severe budget constraints, voted to make the expenditure of state funds on human scoring of assessments illegal.2 More recently, the state has contemplated signing on to the “common core” standards and is monitoring other changes that may become necessary as a result of the Race to the Top Initiative or reauthorization of the federal Elementary and Secondary Education Act.

Mattson reflected on the process so far, noting that despite the examples

1

The technical manual and other information about the program are available at http://education.state.mn.us/MDE/Accountability_Programs/Assessment_and_Testing/Assessments/MCA/TechReports/index.html [April 2010].

2

State money could still be used for machine scoring of assessments. Because Minnesota has no board of education, the legislature is responsible for overseeing the operations of the department of education; federal dollars had been used to support human scoring.

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

of other innovative assessment programs, many aspects of the MCA were new and had to be developed from scratch. These elements included operational issues, such as a means for conveying what was needed to testing contractors, estimating costs, and supporting research and development efforts. They also included fundamental design issues, and Mattson noted that often the content specialists were far ahead of the software designers in conceptualizing what could be done. Technical challenges—from designing a test security protocol to preparing schools to load approximately 300-475 megabytes of test content onto their servers—required both flexibility and patience. A Statewide Assessment Technology Work Group helped the team identify and address many of the technical challenges, and Mattson pointed to this group as a key support.

For Mattson, it was important that the developers were not deterred by the fact that there were no paths to follow in much of development process. The success of the assessment thus far, in his view, has hinged on the fact that the state was able to think ambitiously. The leaders had enthusiastic support from teachers, as well as grant funding and other community support, which allowed them to sustain their focus on the primary goal of developing an assessment that would target the skills and knowledge educators believed was most important. The flexibility that has also been a feature of the MCA since the beginning—the state assessment staff’s commitment to working with and learning from all of the constituencies concerned with the results—should allow them to successfully adapt to future challenges, Mattson said.

POLICY IMPLICATIONS

The three state examples, suggested discussant Lorraine McDonnell, highlight the familiar tension between the “missionaries” who play the important role of seeking ways to improve the status quo and those who raise the sometimes troublesome questions about whether a proposed solution addresses the right problem, whether the expected benefits will outweigh the costs, and whether the innovation can be feasibly implemented. She distilled several policy lessons from the presentations.

First, for an assessment to be successful, it is clear that the testing technology has to be well matched to the policy goals the assessment is intended to serve. Accurate educational measurement may be a technical challenge, but assessment policy cannot be clearly understood independent of its political function. Whether the function is to serve as a student- or school-level accountability device, to support comparisons across schools or jurisdictions, or to influence the content and mode of classroom instruction, what is most important is to ensure that the goals are explicitly articulated and agreed on. McDonnell observed that in many states test developers and politicians had not viewed the function of the state assessment in the same way. As a result, test developers could not meet the policy makers’ expectations, and support for the

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

assessment weakened. When policy makers expect the test to serve multiple purposes, the problem is most acute—and policy makers may not agree among themselves about an assessment’s function.

It is also very important that the testing system be integrated into the broader instructional system, McDonnell advised. This idea was an element of the argument for systemic reform that was first proposed during the 1990s (Smith and O’Day, 1991) and has been prominent in reform rhetoric, but it has not played a major role in the current common standards movement. She pointed out that although the conditions that truly support effective teaching and learning should be central, they “appear once again to have been relegated to an afterthought.” Support is needed to strengthen instructional programs, as well as assessment programs. As numerous participants did throughout the workshop, she highlighted the importance of a comprehensive and coherent system of standards, instruction, and assessment.

States and the federal government have tended, she argued, to select instruments that were easy to deploy—such as tests—and to underinvest in measures such as curricula and professional development that could help to build schools’ capability to improve education. Yet unless teachers are provided with substantial opportunities to learn about the deeper curricular implications of innovative assessments and to reshape their instruction in light of that knowledge, the result of any high-stakes assessment is likely to be more superficial test preparation, or what McDonnell called “rubric-driven instruction.” This collision between policy pressure for ambitious improvements in achievement and the weak capability of schools and systems to respond was an enduring dilemma during the first wave of innovation, in the 1990s, and McDonnell believes that it has not been resolved.

Another lesson is that policy makers and test designers need to understand the likely tradeoffs associated with different types of assessments and then decide which goals they want to advance and which ones they are willing to forgo or minimize. The most evident tension is between using tests as accountability measures and using them as a way to guide and improve instruction. As earlier workshop discussions showed, McDonnell said, these two goals are not necessarily mutually exclusive, but pursuing both with a single instrument is likely to make it more difficult to obtain high-quality results.

This lesson relates to another, which may be obvious, McDonnell suggested, but appears to be easily overlooked: that if a process is to be successful, every constituency that will be affected by an assessment must have ample opportunity to participate throughout it. The differing perspectives of psychometricians and curriculum developers, for example, need to be reconciled if an assessment system is to be successful, but parents, teachers, and other interests need to be heard as well. If developers fail to fully understand and take into account public attitudes, they may encounter unexpected opposition to their plans. Conveying the rationale for an assessment approach, and the expected

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×

benefits and costs, to policy makers and the public may require careful planning and concerted effort. It is often at the district level that the critical communications take place, and too often, McDonnell said, district leaders have been left unprepared for this important aspect of the process. The benefit of clear and thorough communication is that stakeholders are more likely to continue to support a program through technical difficulties, if they have a clear understanding of the overall goals.

Finally, McDonnell stressed, people need to remember that the implementation of innovative assessments takes time. It is very important to build in an adequate development cycle that allows for a gradual, phased roll-out and for adaptation to potential problems. In several of the experiences discussed at the workshop, rushed implementation led to technical problems, undue stress on teachers and students, and a focus on testing formats at the expense of clear connections to curriculum. In several states testing experts acquiesced to political pressure to move quickly in a direction that the testing technology could not sustain. Programs that have implemented innovative features gradually, without dismantling the existing system, have had more flexibility to adapt and learn from experience.

These policy lessons, as well as a growing base of technical advances, can be very valuable for today’s “missionaries,” McDonnell observed. However, although past experience provides lessons, it may also have left a legacy of skepticism among those who had to deal with what were in some cases very disruptive experiences. Fiscal constraints are also likely to be a problem for some time to come, and it is not clear that states will be able to sustain new forms of assessment that may be more expensive after initial seed funding is exhausted. She also noted that the common standards movement and the Race to the Top Initiative have not yet become the focus of significant public attention, and there is no way to predict whether they will become the objects of ideological controversies, as have past education reforms. None of these are reasons not to experiment with new forms of assessment, McDonnell concluded, but “they are reasons for going about the enterprise in a technically more sophisticated way than was done in the past and to do it with greater political sensitivity and skill.”

Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 41
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 42
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 43
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 44
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 45
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 46
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 47
Suggested Citation:"4 Political Experiences and Considerations." National Research Council. 2010. Best Practices for State Assessment Systems, Part I: Summary of a Workshop. Washington, DC: The National Academies Press. doi: 10.17226/12906.
×
Page 48
Next: 5 Opportunities for Better Assessment »
Best Practices for State Assessment Systems, Part I: Summary of a Workshop Get This Book
×
Buy Paperback | $32.00 Buy Ebook | $25.99
MyNAP members save 10% online.
Login or Register to save!
Download Free PDF

Educators and policy makers in the United States have relied on tests to measure educational progress for more than 150 years. During the twentieth century, technical advances, such as machines for automatic scoring and computer-based scoring and reporting, have supported states in a growing reliance on standardized tests for statewide accountability.

State assessment data have been cited as evidence for claims about many achievements of public education, and the tests have also been blamed for significant failings. As standards come under new scrutiny, so, too, do the assessments that measure their results. The goal for this workshop, the first of two, was to collect information and perspectives on assessment that could be of use to state officials and others as they review current assessment practices and consider improvements.

  1. ×

    Welcome to OpenBook!

    You're looking at OpenBook, NAP.edu's online reading room since 1999. Based on feedback from you, our users, we've made some improvements that make it easier than ever to read thousands of publications on our website.

    Do you want to take a quick tour of the OpenBook's features?

    No Thanks Take a Tour »
  2. ×

    Show this book's table of contents, where you can jump to any chapter by name.

    « Back Next »
  3. ×

    ...or use these buttons to go back to the previous chapter or skip to the next one.

    « Back Next »
  4. ×

    Jump up to the previous page or down to the next one. Also, you can type in a page number and press Enter to go directly to that page in the book.

    « Back Next »
  5. ×

    Switch between the Original Pages, where you can read the report as it appeared in print, and Text Pages for the web version, where you can highlight and search the text.

    « Back Next »
  6. ×

    To search the entire text of this book, type in your search term here and press Enter.

    « Back Next »
  7. ×

    Share a link to this book page on your preferred social network or via email.

    « Back Next »
  8. ×

    View our suggested citation for this chapter.

    « Back Next »
  9. ×

    Ready to take your reading offline? Click here to buy this book in print or download it as a free PDF, if available.

    « Back Next »
Stay Connected!