Skip to main content

Study Shows ‘Readability’ Scores Are Largely Inaccurate

Teachers, parents and textbook companies use technical “readability” formulas to determine how difficult reading materials are and to set reading levels by age group. But new research from North Carolina State University shows that the readability formulas are usually inaccurate and offer little insight into which age groups will be able to read and understand a text.

“Teachers often use readability levels when giving reading assignments to students,” says Dr. John Begeny, an associate professor of psychology at NC State and lead author of a paper describing the work. “We wanted to know if the readability formulas are valid, or if teachers who think they’re assigning a simpler book to struggling readers, for example, may actually be assigning a more difficult one.”

For the study, researchers had 360 students – ranging from second to fifth grade – read six written passages out loud. The researchers assessed the students’ performance, giving each student an “oral reading fluency” score, which is considered a good metric for measuring reading ability.

The researchers then used eight different readability formulas to see which level each formula gave to the six written passages. Results varied widely, with one passage being rated from first grade to fifth grade level.

The levels assigned by the readability formulas were then compared with researchers’ assessments of each student’s actual ability to read the material. Seven of the eight readability formulas were less than 49 percent accurate, with the worst formula scoring only 17 percent accuracy. The highest-rated formula was accurate 79 percent of the time.

“Overall, this work shows that teachers and parents should be very cautious about using readability levels when giving reading assignments to students,” Begeny says.

The paper, “Can Readability Formulas Be Used to Successfully Gauge Difficulty of Reading Materials?” is published in the January issue of the journal Psychology in the Schools. The paper was co-authored by Diana Greene, a former graduate student at NC State.

-shipman-

Note to Editors: The study abstract follows.

“Can Readability Formulas Be Used to Successfully Gauge Difficulty of Reading Materials?”

Authors: John C. Begeny and Diana J. Greene, North Carolina State University

Published: January 2014, Psychology in the Schools

DOI: 10.1002/pits.21740

Abstract: A grade level of reading material is commonly estimated using one or more readability formulas, which purport to measure text difficulty based on specified text characteristics. However, there is limited direction for teachers and publishers regarding which readability formulas (if any) are appropriate indicators of actual text difficulty. Because oral reading fluency (ORF) is considered one primary indicator of an elementary aged student’s overall reading ability, the purpose of this study was to assess the link between leveled reading passages and students’ actual ORF rates. ORF rates of 360 elementary-aged students were used to determine whether reading passages at varying grade levels are, as would be predicted by readability levels, more or less difficult for students to read. Results showed that a small number of readability formulas were fairly good indicators of text, but this was only true at particular grade levels. Additionally, most of the readability formulas were more accurate for higher ability readers. One implication of the findings suggests that teachers should be cautious when making instructional decisions based on purported “grade-leveled” text, and educational researchers and practitioners should strive to assess difficulty of text materials beyond simply using a readability formula.