Explaining reading variance by student subgroup: should we move beyond oral reading fluency?
Background Studies have demonstrated strong predictive and concurrent validity of curriculum‐based measures (CBMs) with high‐stakes reading tests, mainly using oral reading fluency (ORF) to predict outcomes. Some studies have found that CBMs' predictive accuracy might vary across demographic su...
Gespeichert in:
Veröffentlicht in: | Journal of research in reading 2021-11, Vol.44 (4), p.757-786 |
---|---|
Hauptverfasser: | , , |
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Background
Studies have demonstrated strong predictive and concurrent validity of curriculum‐based measures (CBMs) with high‐stakes reading tests, mainly using oral reading fluency (ORF) to predict outcomes. Some studies have found that CBMs' predictive accuracy might vary across demographic subgroups of students. In this study, we investigate whether additional CBMs of reading comprehension account for variance beyond ORF and whether these reading variables predict reading comprehension differently for demographic subgroups.
Methods
Curriculum‐based measures were administered to 9,967 students at the end of Grade 3 in the United States, concurrent with a high‐stakes standardised state reading comprehension achievement test (RCAT). Hierarchical regression models with RCAT as the outcome were used to test (a) whether CBMs of reading comprehension accounted for significant variance beyond ORF and (b) if the reading variables differentially predicted RCAT for different subgroups of students.
Results
Oral reading fluency explained significant variation in RCAT, and the addition of comprehension CBMs accounted for small amounts of unique variance. The contribution of ORF in explaining variation in reading comprehension varied by demographic subgroups.
Conclusions
The contributions of comprehension CBMs in predicting high‐stakes standardised reading comprehension tests were not meaningful, only accounting for small amounts of unique variance. Additionally, the contribution of ORF in explaining variation in the high‐stakes standardised reading comprehension test varied by student demographics, likely due to complex reasons specific to those individual factors.
Highlights
What is already known about this topic
CBMs of ORF account for significant variance in reading comprehension achievement on high‐stakes standardised reading tests.
Failure to include some measure of comprehension in CBMs could be problematic because fluency does not ensure comprehension. When comprehension CBMs have been included, findings have been mixed.
Some studies have pointed to differences in how well ORF predicts reading comprehension for various subgroups of students.
What this paper adds
We examine the contribution of comprehension CBMs over and above ORF in predicting student performance on a high‐stakes reading comprehension test using a large, diverse sample in the United States.
We replicate others' findings on the contribution of ORF in predicting variance in students' high‐stakes |
---|---|
ISSN: | 0141-0423 1467-9817 |
DOI: | 10.1111/1467-9817.12371 |