Explaining reading variance by student subgroup: should we move beyond oral reading fluency?

Background Studies have demonstrated strong predictive and concurrent validity of curriculum‐based measures (CBMs) with high‐stakes reading tests, mainly using oral reading fluency (ORF) to predict outcomes. Some studies have found that CBMs' predictive accuracy might vary across demographic su...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:Journal of research in reading 2021-11, Vol.44 (4), p.757-786
Hauptverfasser: Amendum, Steven J., Conradi Smith, Kristin, Liebfreund, Meghan D.
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Background Studies have demonstrated strong predictive and concurrent validity of curriculum‐based measures (CBMs) with high‐stakes reading tests, mainly using oral reading fluency (ORF) to predict outcomes. Some studies have found that CBMs' predictive accuracy might vary across demographic subgroups of students. In this study, we investigate whether additional CBMs of reading comprehension account for variance beyond ORF and whether these reading variables predict reading comprehension differently for demographic subgroups. Methods Curriculum‐based measures were administered to 9,967 students at the end of Grade 3 in the United States, concurrent with a high‐stakes standardised state reading comprehension achievement test (RCAT). Hierarchical regression models with RCAT as the outcome were used to test (a) whether CBMs of reading comprehension accounted for significant variance beyond ORF and (b) if the reading variables differentially predicted RCAT for different subgroups of students. Results Oral reading fluency explained significant variation in RCAT, and the addition of comprehension CBMs accounted for small amounts of unique variance. The contribution of ORF in explaining variation in reading comprehension varied by demographic subgroups. Conclusions The contributions of comprehension CBMs in predicting high‐stakes standardised reading comprehension tests were not meaningful, only accounting for small amounts of unique variance. Additionally, the contribution of ORF in explaining variation in the high‐stakes standardised reading comprehension test varied by student demographics, likely due to complex reasons specific to those individual factors. Highlights What is already known about this topic CBMs of ORF account for significant variance in reading comprehension achievement on high‐stakes standardised reading tests. Failure to include some measure of comprehension in CBMs could be problematic because fluency does not ensure comprehension. When comprehension CBMs have been included, findings have been mixed. Some studies have pointed to differences in how well ORF predicts reading comprehension for various subgroups of students. What this paper adds We examine the contribution of comprehension CBMs over and above ORF in predicting student performance on a high‐stakes reading comprehension test using a large, diverse sample in the United States. We replicate others' findings on the contribution of ORF in predicting variance in students' high‐stakes
ISSN:0141-0423
1467-9817
DOI:10.1111/1467-9817.12371