69ý

Assessment

New Details Surface About Common Assessments

By Catherine Gewertz — January 10, 2012 6 min read
  • Save to favorites
  • Print
Email Copy URL

With one set of academic standards now serving as the educational guideposts in nearly every state, questions are hovering about what the tests for those standards will look like. But gradually, details are emerging that show plans that could fundamentally change the U.S. testing landscape.

Documents issued by the two groups of states that are designing the tests show that they seek to harness the power of computers in new ways and assess skills that multiple-choice tests cannot. Those plans are very fluid, however, since several years of design, dialogue, revision, piloting, and reworking lie ahead before the assessments are ready in 2014-15. But early documents offer glimpses of the groups’ thinking.

“This stuff is a very big deal, and it’s a huge departure from the kinds of tests most kids currently take,” said Chuck Pack, a national-board-certified math teacher at Tahlequah High School, in Tahlequah, Okla., a small town outside Tulsa.

“As classroom teachers, we’re sitting here waiting to know what our kids are going to be expected to do. We have the standards—what they’re supposed to know—but now how are they supposed to be able to demonstrate that? Documents like this help us get our heads around that,” said Mr. Pack, who serves on an advisory board that is guiding CTB/McGraw-Hill as it designs “next generation” assessments.

The information is trickling out in solicitations issued in the past two months by the SMARTER Balanced Assessment Consortium and the Partnership for Assessment of Readiness for College and Careers, or PARCC, for vendors to work on the tests. Those two consortia of states are using $360 million in federal Race to the Top money to create new assessments for the Common Core State Standards, which all but four states have adopted.

“Every major publisher that has been a state assessment contractor in the past, and many others with an educational reform orientation, are paying attention” to those solicitations, said Alan J. Theimann, the legislative counsel for the Association of Test Publishers.

A Dec. 30 solicitation by PARCC, seeking vendors to write test items, describes the consortium’s vision of its testing system in more detail than did previous documents. It expects to award that contract in April to “multiple” vendors to design half the test items, and renew the contract to some of those vendors to craft the rest.

The solicitation covers the development of the two pieces of the test that will yield students’ summative scores in mathematics and English/language arts and be used for accountability purposes: a computer-based end-of-year test and a performance-based assessment given toward the end of the year. The scope of work also includes developing midyear formative assessments that are part of PARCC’s system but are optional for states.

Digging Into Text

A preliminary blueprint of PARCC’s English/language arts exam shows that the performance-based assessment, spread over two days, would involve a “research simulation” that asks students to read a suite of texts, including an “anchor” text such as a speech by a prominent historical figure. They would have to answer questions that require them to cite evidence from the text for their answers and write an essay. Another aspect of the performance-based test would require students to “engage” with literature (grades 3-5) or conduct literary analysis (grades 6-11) using a combination of shorter and longer texts.

The end-of-year exam would employ six literary and informational texts and ask students to respond to machine-scorable questions, including ones that demand comparison and synthesis of the readings.

The end-of-year test in English/language arts would yield at least half of a student’s points in that topic. One-third to one-half would come from the performance-based test, according to the preliminary blueprint.

PARCC’s math test will include three types of questions: “innovative,” machine-scorable, computer-based items; items that call for written arguments or justifications; critiques of mathematical reasoning, or proof that students “attended to precision” in math; and items involving real-world scenarios. The performance-based assessment in math will count for 40 percent to 50 percent of a student’s points in that subject, and the end-of-course exam will yield 50 percent to 60 percent of the points.

The math exams will focus on solving problems in the “major content areas” at each grade level, as well as demonstrating conceptual understanding, fluency and mathematical reasoning, and applying knowledge to real-world problems.

At the high school level, PARCC will develop two series of end-of-course math tests: a traditional one—Algebra 1, geometry, and Algebra 2—and one that integrates those topics. Those parallel pathways reflect choices educators can make about how to design math courses from the common standards.

The solicitation document answers a question that had been circulating among some educators of young children. PARCC said that its tests will be given by computer to students in grades 6-11, but those in grades 3-5 will answer questions with pencil and paper because of concerns about younger children’s keyboarding skills.

Teacher Participation

Documents issued recently by the SMARTER Balanced consortium offer a less-descriptive preview of its tests, largely because work on an earlier solicitation, to design item specifications, isn’t yet complete, and informs other parts of the test design. That request for proposals, issued in July, and its content specifications, released in August, represent the most detailed version of the consortium’s ideas. (“Consortia Flesh Out Concepts for Common Assessments,” Aug. 24, 2011.)

In a request for proposals issued last month, SMARTER Balanced seeks development of 10,000 selected-response or constructed-response items and 420 performance tasks in math and English/language arts to facilitate pilot-testing in the 2012-13 school year. Most will be scored by machine, the document says.

Part of the work will be conducting research to find out which types of items are best suited to automated scoring and which must be scored by hand.

The request for proposals also asks the prospective vendor to hire and train teachers from SMARTER Balanced states to write items and tasks and review items for content alignment, accessibility, and bias. PARCC documents say that teachers will help shape the tests by serving on local committees reviewing test items. They will also be involved in developing model instructional units, diagnostic assessments, professional-development modules, and other PARCC resources.

In September, SMARTER Balanced issued a solicitation for development of guidelines for accessibility and accommodations for English learners and students with disabilities. PARCC plans such a solicitation this year, as well as requests for work on other parts of its testing system, such as its early-year diagnostic assessments and tests of speaking and listening skills.

PARCC has contracted with the Dana Center at the University of Texas at Austin to build prototype assessment tasks in math, and with the University of Pittsburgh’s Institute for Learning to generate such items in literacy. Those items are slated for release this summer.

A version of this article appeared in the January 11, 2012 edition of Education Week as New Details Surface About Common State Assessments

Events

This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Artificial Intelligence Webinar
AI and Educational Leadership: Driving Innovation and Equity
Discover how to leverage AI to transform teaching, leadership, and administration. Network with experts and learn practical strategies.
Content provided by 
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
School Climate & Safety Webinar
Investing in Success: Leading a Culture of Safety and Support
Content provided by 
Assessment K-12 Essentials Forum Making Competency-Based Learning a Reality
Join this free virtual event to hear from educators and experts working to implement competency-based education.

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.

Read Next

Assessment Opinion 69ý Shouldn't Have to Pass a State Test to Graduate High School
There are better ways than high-stakes tests to think about whether students are prepared for their next step, writes a former high school teacher.
Alex Green
4 min read
Reaching hands from The Creation of Adam of Michelangelo illustration representing the creation or origins of of high stakes testing.
Frances Coch/iStock + Education Week
Assessment Opinion Why Are Advanced Placement Scores Suddenly So High?
In 2024, nearly three-quarters of students passed the AP U.S. History exam, compared with less than half in 2022.
10 min read
Image shows a multi-tailed arrow hitting the bullseye of a target.
DigitalVision Vectors/Getty
Assessment Grades and Standardized Test Scores Aren't Matching Up. Here's Why
Researchers have found discrepancies between student grades and their scores on standardized tests such as the SAT and ACT.
5 min read
Student writing at a desk balancing on a scale. Weighing test scores against grades.
Vanessa Solis/Education Week + Getty Images
Assessment Why Are States So Slow to Release Test Scores?
Nearly a dozen states still haven't put out scores from spring tests. What's taking so long?
7 min read
Illustration of a man near a sheet of paper with test scores on which lies a magnifying glass and next to it is a question mark.
iStock/Getty