69ý

Assessment

More Errors Are Seen In the Scoring of Tests, Boston Researchers Say

By Michelle Galley — June 18, 2003 4 min read
  • Save to favorites
  • Print
Email Copy URL

right The report, is available from . (Requires .)

The number of human errors identified in standardized-test results has risen dramatically in recent years, calling into question whether such exams should be the only measure for such high-stakes decisions as school ranking, student graduation, and promotion to the next grade, says a report that examines those mistakes.

The fallout from inaccurate scores can mean that a qualified teacher is not certified, a competent student is not allowed to graduate, or a literate 3rd grader cannot enter 4th grade with her peers, said Kathleen Rhoades, an author of the report.

Released June 5 by the independent National Board for Educational Testing and Public Policy, based at Boston College’s school of education, the study found that on all standardized tests given in 1976, only one mistake by the testing industry was reported. That number jumped to 14 in 1999 and has stayed relatively steady since then, the report says.

Ms. Rhoades and co-author George Madaus, a senior fellow with the National Board on Educational Testing and Public Policy, worry that the numbers will climb higher as states scramble to satisfy the accountability demands in the “No Child Left Behind” Act of 2001.

Between the lack of oversight in the testing industry and the increased demands that will be placed on testing companies as a result of the federal law, the future “is ripe for the proliferation of undetected human error in educational testing,” the report warns.

More Tests, More Scrutiny

One reason for the apparent increase could simply be that more testing is taking place now, said Collin Earnst, a spokesman for the Boston-based Houghton Mifflin Co., one of the nation’s leading test publishers. “If we add another 5 million cars to the road, will we expect more car accidents? We probably would,” he said.

The proliferation of testing is not the only factor, however, according to Ms. Rhoades, a graduate research assistant at the education school. Test scores are scrutinized much more carefully now, she said, because they are used so prominently to make important decisions.

Using newspaper accounts and past research, the Boston College report chronicles some recent blunders made in the testing industry, and their dramatic effects.

For example, in 1999, a statistician with the Tennessee education department noticed that two-thirds of the state’s scores on the TerraNova test had decreased, and asked the testing company, CTB/McGraw-Hill, to review the scores, according to the report.

Eventually, the company discovered that “a programming error caused the percentile rankings on the TerraNova to be too low at the low end of the scale and too high at the high end of the scale,” the report says.

That lapse led to incorrect scores for roughly 250,000 students in six states.

Michael H. Kean, the vice president of public and governmental affairs for the Monterey, Calif.-based CTB/McGraw-Hill, said he had not read the report and thus declined to comment.

‘Caveat Emptor’

It is not always easy to examine data from testing companies closely, the report says. The industry itself is “shrouded in secrecy,” and testing companies “often classify their information as confidential and release only bits,” the report maintains.

Houghton Mifflin’s Mr. Earnst disagreed. “Everything in this industry of educational testing is very visible,” including test scores and the tests that are given, he said. “A lot of the processes for how testing companies are selected are public record.”

Many school districts and state education departments select testing companies based on their past performance, said Lee Baldwin, the president of the National Association of Test Directors.

“What you depend on with a contractor is that they are going to give you scores that are accurate,” he said. “If they don’t, it raises serious questions with their credibility.”

Mistakes come with the testing territory, he added. “No system is perfect; anything we do is subject to error,” said Mr. Baldwin, who is also the senior director of student assessment and program evaluation for the 160,000-student Orange County schools in Florida.

Still, districts and state education departments should minimize errors by implementing “careful procedures” on how tests are scored and how the outcomes are recorded and measured, he said.

Passage of the No Child Left Behind law, which requires annual testing in grades 3-8 and once during the high school years, makes such precautions more imperative, Mr. Baldwin said. The testing industry is rapidly expanding, he pointed out, and because there are no industry regulations, “almost anyone can claim to be a test publisher and test scorer.”

Consequently, Mr. Baldwin says: “Caveat emptor,” or let the buyer beware.

Related Tags:

Events

This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
Artificial Intelligence Webinar
AI and Educational Leadership: Driving Innovation and Equity
Discover how to leverage AI to transform teaching, leadership, and administration. Network with experts and learn practical strategies.
Content provided by 
This content is provided by our sponsor. It is not written by and does not necessarily reflect the views of Education Week's editorial staff.
Sponsor
School Climate & Safety Webinar
Investing in Success: Leading a Culture of Safety and Support
Content provided by 
Assessment K-12 Essentials Forum Making Competency-Based Learning a Reality
Join this free virtual event to hear from educators and experts working to implement competency-based education.

EdWeek Top School Jobs

Teacher Jobs
Search over ten thousand teaching jobs nationwide — elementary, middle, high school and more.
Principal Jobs
Find hundreds of jobs for principals, assistant principals, and other school leadership roles.
Administrator Jobs
Over a thousand district-level jobs: superintendents, directors, more.
Support Staff Jobs
Search thousands of jobs, from paraprofessionals to counselors and more.

Read Next

Assessment Opinion 69ý Shouldn't Have to Pass a State Test to Graduate High School
There are better ways than high-stakes tests to think about whether students are prepared for their next step, writes a former high school teacher.
Alex Green
4 min read
Reaching hands from The Creation of Adam of Michelangelo illustration representing the creation or origins of of high stakes testing.
Frances Coch/iStock + Education Week
Assessment Opinion Why Are Advanced Placement Scores Suddenly So High?
In 2024, nearly three-quarters of students passed the AP U.S. History exam, compared with less than half in 2022.
10 min read
Image shows a multi-tailed arrow hitting the bullseye of a target.
DigitalVision Vectors/Getty
Assessment Grades and Standardized Test Scores Aren't Matching Up. Here's Why
Researchers have found discrepancies between student grades and their scores on standardized tests such as the SAT and ACT.
5 min read
Student writing at a desk balancing on a scale. Weighing test scores against grades.
Vanessa Solis/Education Week + Getty Images
Assessment Why Are States So Slow to Release Test Scores?
Nearly a dozen states still haven't put out scores from spring tests. What's taking so long?
7 min read
Illustration of a man near a sheet of paper with test scores on which lies a magnifying glass and next to it is a question mark.
iStock/Getty