Professional Documents
Culture Documents
Using Small Group Instruction To Improve Students Reading Fluency An Evaluation of The Existing Research
Using Small Group Instruction To Improve Students Reading Fluency An Evaluation of The Existing Research
Using Small Group Instruction To Improve Students Reading Fluency An Evaluation of The Existing Research
To cite this article: John C. Begeny, Rebecca A. Levy & Stacey A. Field (2018) Using Small-Group
Instruction to Improve Students' Reading Fluency: An Evaluation of the Existing Research, Journal
of Applied School Psychology, 34:1, 36-64, DOI: 10.1080/15377903.2017.1328628
The National Reading Panel (National Institute of Child Health and Human
Development, 2000) and several subsequent U.S. panels and publications (e.g.,
Armbruster, Lehr, & Osborne, 2001; Lonigan & Shanahan, 2010) identified oral
reading fluency as one of the five foundational skills crucial for successful reading
achievement. Reading fluency can be defined as the “ability to read connected text
rapidly, smoothly, effortlessly, and automatically with little conscious attention to
the mechanics of reading, such as decoding,” (Meyer & Felton, 1999, p. 284), and
is considered essential for comprehension, generalization, and maintenance of all
reading skills (Bonfiglio, Daly, Martens, Lin, & Corsaut, 2004; Daane, Campbell,
Grigg, Goodman, & Oranje, 2005; Fuchs, Fuchs, Hosp, & Jenkins, 2001; Torgesen
& Hudson, 2006).
Unfortunately, far too many students have inadequate reading fluency skills. For
example, Snow, Burns, and Griffin (1998) argued that 25–40% of students’ academic
careers are at risk because they are not reading quickly or easily enough, and nation-
ally representative studies of students’ reading fluency estimate approximately 40%
CONTACT John C. Begeny john_begeny@ncsu.edu Department of Psychology, North Carolina State University,
Campus Box , Raleigh, NC –, USA.
© Taylor & Francis Group, LLC
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 37
of U.S. fourth-grade students are nonfluent readers (Daane et al., 2005). Given these
data and the importance of fluency for reading comprehension (Daane et al., 2005;
Fuchs et al., 2001), it is not surprising that 31% of U.S. fourth-grade students read
below a basic level and 64% of them read below a proficient level (National Assess-
ment of Educational Progress, 2015).
Small-group instruction
Method
Inclusion criteria
Given our focus on reviewing the research on SG interventions designed specifically
to target reading fluency, a research report was included if it was published prior to
2016 and met each of the following inclusion criteria.
r The research was reported in English within a peer-reviewed journal or book
chapter.
r The report presented previously unpublished findings.
r The study was experimental or quasiexperimental and generated quantitative
outcomes.
r The report described a reading intervention that was implemented by an adult
with three or more children at the same time.
r The study included at least one dependent variable that measured students’
reading fluency.
r The author(s) of the study indicated (explicitly or implicitly) that the primary
goal of the intervention was to improve students’ reading fluency. However, if
reading fluency was just one area being measured in the study and the authors
did not state that the main goal of the intervention was to improve fluency,
it was not included because we were interested in reviewing targeted (rather
than broad based) interventions for reading fluency. This criterion is consistent
with our study goals and the fact that many students struggle specifically with
reading fluency and would benefit from a targeted intervention (e.g., Denton
et. al., 2011; Fletcher et al., 2011).
r The SG instruction described in the study included at least two strategies (e.g.,
repeated readings, modeling, reward procedures, performance feedback) that
have been shown in prior research to improve reading fluency. This criterion
helped to ensure that the goal of the intervention was to improve participants’
reading fluency.
Database searches
Using the PsycINFO and ERIC database systems, a computer search of the literature
was conducted using combinations of unique keywords. Specifically, the following
22 separate searches were conducted: small group and fluency, small-group and flu-
ency, instruction and group and fluency, grouping and fluency, assisted reading and
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 41
fluency and group, RTI and fluency, response to intervention and fluency, response-to-
intervention and fluency, Tier 2 and fluency, Tier 3 and fluency, group and repeated
reading and fluency, group and model reading and fluency, group and repeated reading
and automaticity, fluency and Read Naturally, fluency and Reader’s Theater, fluency
and Great Leaps, fluency and RAVE-O, fluency and everybody reads, fluency and Fast
Start, fluency and HELPS program, fluency and Fast for Word, reading fluency and
the fluency assessment system. Results from this search yielded 553 unique studies,
which were reviewed closely by a member of the research team, Field, who has con-
siderable experience with the topic of our investigation and the process of including
or excluding studies according to established literature review criteria. As part of the
process, the full article or abstract were reviewed for potential inclusion and all arti-
cles that clearly did not meet the inclusion criteria were excluded. The articles that
could potentially meet the inclusion criteria (n = 12) were independently reviewed
by the entire research team. With 100% agreement, nine studies met the inclusion
criteria from these search procedures.
Next, we searched Google Scholar with the terms small group AND reading flu-
ency instruction as a means to confirm the comprehensiveness of our original search
and determine if there were additional articles that would meet the inclusion crite-
ria. Because this search resulted in 117,000 hits (which would be expected given the
nature of the Google search engine and our search terms), and because we knew the
vast majority of these hits would not be relevant to our study or redundant with prior
searches, the author, Field, who handled the initial search of articles from PsycINFO
and ERIC carefully reviewed the first 500 Google Scholar hits. No additional studies
met the inclusion criteria based on the search with Google Scholar.
Hand searches
After the database searches, each member of the research team conducted a hand
search of each study that met the inclusion criteria. This was done to determine if
other references were made within the source to other potentially eligible articles or
chapters not already obtained or reviewed. For the same purpose, we also read each
of the six published meta-analyses (Chard et al., 2002; Kuhn & Stahl, 2003; Meyer
& Felton, 1999; Morgan & Sideridis, 2006; National Institute of Child Health and
Human Development, 2000; Therrien, 2004) that were available at the time of the
search and were on the topic of reading fluency instruction. The References sections
of those studies were also carefully reviewed. If potential sources were referenced
in the hand search procedures, we obtained the potentially includable source and
reviewed it for possible inclusion. Using hand-search procedures, and with 100%
agreement among those conducting the search, two additional studies were identi-
fied as meeting the inclusion criteria.
we found one or more previously published articles from the author, referenced the
author’s work that we identified for inclusion, and asked if he or she previously pub-
lished or had a article in press that might also meet our inclusion criteria. As needed,
one follow-up email was sent to reiterate our request. Of the five authors contacted,
everyone replied and we identified one additional article that met our inclusion cri-
teria (i.e., an article that was in press at the time of our search). Based on all of our
search procedures, 12 articles met the inclusion criteria (representing 13 separate
studies) and each study was summarized and coded.
To ensure reliable summaries and coding of the included studies, two authors,
Begeny and Field, independently coded 100% of the articles and a third author,
Levy, reviewed their coding to assess agreements and disagreements. All disagree-
ments were discussed by the entire research team and full consensus was derived
about how to accurately code the information, and the agreed-on coding (with 100%
agreement) is reflected within this report. Coding occurred in two phases. First, to
answer Research Questions 1, 2, and 4, we coded all of the information listed in
Table 1 (with data from this coding summarized in Tables 2–4). Across all categories
and studies, the initial percentage agreement between the two coders was 88.8%, and
then a consensus model among all authors was used to resolve disagreements (i.e.,
all authors discussed each disagreement until full consensus was achieved on the
respective issue).
To answer Research Question 3, the second phase of coding was completed to
assess the methodological quality of the included studies and determine to what
extent SG instruction practices that target reading fluency can be considered an
evidence-based practice (EBP). Across all categories and studies, the initial percent-
age agreement between the two coders was 97.3%, and then a consensus model was
used to resolve disagreements. Tables 5 and 6 summarize the coding categories and
data related to this phase of coding. The categories listed in the left-hand column
of those tables come directly from the criteria described by Horner et al. (2005) for
SCDs and Gersten et al. (2005) for GDS. We followed the definitions and criteria
from those resources, and used the binary coding scheme of met or did not meet
reported in a recent literature review by Sreckovic, Common, Knowles, and Lane
(2014). Due to space restrictions, we refer the reader to those three resources for
more description of each coding category.
In this study we followed the coding rules described by (Sreckovic et al., 2014),
with only two exceptions. First, in determining whether dependent measures were
reliable and valid, we did not require the study to specifically provide reliability and
validity estimates for measures that have been described in numerous past studies
and by national panels as having adequate evidence of reliability and validity. For
example, with words read correctly per minute (WCPM) via curriculum-based mea-
sures of oral reading fluency (CBM-R), nearly all studies reported using this depen-
dent variable; but very few studies specifically discussed its reliability and validity.
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 43
Note. : = one-on-one; ELL = English language learner; Ix = intervention; SG = small group.
a Categories presented in Table .
b Categories presented in Table .
c Categories presented in Table .
Instead, most studies referred the reader to resources describing its reliability and
validity. We therefore did not require specific descriptions of reliability or validity
for measures such as these because we know reviewers often request that psychome-
tric information for measures widely used in research be deleted for space purposes.
In fact, one reviewed study (Begeny, Braun, Lynch, Ramsay, & Wendt, 2012) specif-
ically stated that page restrictions prohibited a report of psychometric properties
for the WCPM measure. Second, Gersten et al. (2005) listed one of the “desirable
quality indicators” as: “Did the research report include actual audio or videotape
excerpts that capture the nature of the intervention?” (p. 152). Our interpretation
of that indicator is that the report provides detail or resources specifying the nature
of the intervention. As such, if a study included actual implementation protocols
44
Table . (Continued)
Received Evaluated Small group Grade Screeners Other
Author Year N Demographics Ix : Ix? size (reading level) (rule-out conditions) characteristics ELL (n)
J. C. BEGENY ET AL.
Begeny, Braun, F= N and and (at risk) WCPM (just prior to Urban school in N
Lyncha , M= intervention, all Southeastern
Ramsaya , & B/AA = participants scored United States
Wendt H= in the “at risk”
W= category on the
DIBELS universal
screening measure)
Begeny, Yeager, & F= Y (at risk) WCPM (th–th Urban school in San N
Martínez M= percentile on Jose, Costa Rica
B/AA = Spanish WCPM)
H=
W=
Ross & Begeny F= Y (at risk) WCPM (th–th Three classrooms in a N
M= percentile on rural school in
B/AA = WCPM) Southeastern
H= United States
W=
Note. B/AA = Black or African American; CTOPP = Subtest(s) from the Comprehensive Test of Phonological Processing; F = female; H = Hispanic or Latino; M = male; N = no; NR = not reported;
TOWRE = Test of Word Reading Efficiency; W = White; WCPM = words read correctly per minute (almost always as part of a seasonal benchmark assessment); WJ-III Reading = Reading subtest(s)
from the Woodcock-Johnson III; Y = yes;
a Author not affiliated with university.
Table . Summary of the findings: Characteristics of interventions, interventionists, and experimental design.
Training Implementation Components Manualized Session Session
Author Year Interventionists level integritya Design (n) Ix? Passagesb duration frequency
Kuhn Research team NR NR Between subjects LPP, RR, PF, CR, N Trade books – min x/week
(quasiexperimental) VC ()
Begeny & Martens Research team Trained to mastery by –% of sessions; Multiple baseline WLT, PD, LPP, RR, N Silver, Burdett, and – min NR
researchers mean integrity > M/R () Ginn reading
% series
Begeny & Silber Research team Trained to mastery by % of sessions; mean Alternating treatments WLT, LPP, RR, N DIBELS – min x/week
researchers integrity = M/R, retell ()
–%
Bonfiglioa , Daly, Research team & Some formal training –% of sessions; Multiple probe across tasks LPP, CR, PD, M/R, N Houghton Mifflin – min x/week
Persampieri, & School staff mean integrity= GS, PF () reading series
Andersen –%
McCurdy, Daly, Research team NR % of sessions; mean Multiple probe across tasks LPP, RR, CR, PD, N Silver, Burdett, and – min NR
Gortmaker, integrity= % M/R () Ginn reading
Bonfiglioa , & series, Houghton
Persampieri Mifflin reading
(Experiment ) series, and
researcher
compiled
McCurdy, Daly, School staff NR % of sessions; mean Adapted alternating CR, PD, LPP, M/R, N Silver, Burdett, and min x/week
Gortmaker, integrity= .% treatments RR () Ginn reading
Bonfiglioa , & series
Persampieri
(Experiment )
(Continued on next page)
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY
47
48
Table . (Continued)
Training Implementation Components Manualized Session Session
Author Year Interventionists level integritya Design (n) Ix? Passagesb duration frequency
Begeny, Krouse, Research team Trained to mastery by % of sessions; mean Alternating treatments RR, LPP, LO, M/R N DIBELS – min x/week
Ross, & Mitchell researchers integrity= % ()
J. C. BEGENY ET AL.
Klubnik & Ardoin Research team Trained to mastery by % of sessions; mean Alternating treatments LPP, RR, PD, M/R N Silver, Burdett, and NR NR
researchers integrity=.% () Ginn reading
series
Begeny, Hawkins, Research team Trained to mastery by –% of sessions; Alternating treatments RR, LPP, PD, retell, N Silver, Burdett, and – min x/week
Krouse, & researchers mean integrity= M/R () Ginn reading
Laugle –% series
Ross & Begeny Research team Trained to mastery by –% of sessions; Alternating treatments LPP, RR, retell, PD, N DIBELS min NR
researchers mean integrity= VI, M/R ()
%
Begeny, Braun, School staff Trained to mastery by % of sessions; mean One-group RR, LPP, PD, VC, Y HELPS Curriculum – min –x/week
Lyncha , researchers integrity= % pretest–posttest GS, PF, M/R,
Ramsaya , & (quasiexperimental) retell ()
Wendt
Begeny, Yeager, & Research team Trained to mastery by –% of sessions; Alternating treatments RR, LPP, PD, M/R, N AIMSweb .– min NR
Martínez researchers mean integrity∼ retell ()
%
Ross & Begeny Research team Trained to mastery by –% of sessions; Alternating treatments LPP, RR, retell, N Silver, Burdett, and – min x/week
researchers mean integrity= M/R, PD, PF () Ginn reading
–% series, and
researcher
compiled
Note. CR = choral reading; DIBELS = Dynamic Indicators of Basic Early Literacy Skills; GS = goal setting; HELPS = Helping Early Literacy with Practice Strategies; Ix = intervention; LO = listening only
(to a passage read aloud); LPP = modeling/listening passage preview; M/R = motivational/reward system; N = no; NR = not reported; PD = phrase-drill error correction; PF = performance feedback;
RR = repeated reading; VC = verbal cueing; VI = vocabulary instruction; WLT = word list training; Y = yes.
a A range is sometimes reported because some studies monitored implementation integrity more/less for some components, conditions, or interventionists (e.g., researcher vs. teacher) and reported the
integrity data specifically by component, condition, or interventionist.
b All selected passages (in each of the studies) were assessed according to students’ instructional level and needs, and in this way they were “leveled” before beginning the study.
Table . Summary of the findings: Study duration, measures, data analytic techniques, and outcomes.
Study Dependent Acceptability Analytic Overall Effects of
Author Year duration measures measures methods outcomes SG vs. :
Kuhn weeks QRI, QRI-II, TOWRE, NAEP NR Interpretations based Based on visual analysis of mean NA
Oral Reading Fluency only on descriptive scores, students in the Ix conditions
Scale statistics (e.g., outperformed the control group;
means); no statistical wide reading and RR improved word
analyses computed recognition, prosody, and WCPM.
Wide reading improved
comprehension
Begeny & Martens – weeks WCPM, Maze Student reported Visual analysisa , paired of (%) of students performed NA
comprehension, word samples t tests, better with Ix compared to a
lists, WJ-III Reading Wilcoxon signed treatment as usual condition
ranks t-test
Begeny & Silber NR WCPM NR Visual analysis of (%) of students performed NA
better with each Ix compared to
control condition; combination of all
components was most effective for
all students
Bonfigliob , Daly, weeks WCPM, Level of Student Teacher reported Visual analysis, effect of (%) of students performed NA
Persampieri, & Engagement size calculations better with each Ix compared to
Andersen control condition; the identified
“effective package” was most
effective for three of the four
students
McCurdy, Daly, NR WCPM NR Visual analysis of (%) of students performed NA
Gortmaker, better with Ix compared to baseline
Bonfigliob , & levels of performance
Persampieri
(Experiment )
McCurdy, Daly, weeks WCPM Teacher reported Visual analysis of (%) of students performed NA
Gortmaker, better with Ix compared to the
Bonfigliob , & control condition
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY
Persampieri
(Experiment )
(Continued on next page)
49
Table . (Continued)
50
Begeny, Krouse, Ross, NR WCPM NR Visual analysis of (%) of students performed NA
& Mitchell better with each of the Ix conditions
compared to control condition; RR
was most effective, followed by LPP
Klubnik & Ardoin weeks WCPM NR Visual analysis of (%) of students performed / = %
better with at least one Ix condition
J. C. BEGENY ET AL.
Note. Ix = intervention; NA = not applicable; NAEP = National Assessment of Educational Progress; NR = not reported; QRI = Qualitative Reading Inventory; PND = percentage of nonoverlapping data;
PT = peer tutoring; RR = repeated reading; SEM = standard error of measurement; SG = small group; TOWRE = Test of Word Reading Efficiency; WCPM = Words read correctly per minute; WJ-III
Reading = Reading subtest(s) from the Woodcock- Johnson III.
a Visual analysis refers to visual analysis procedures that are commonly used with single-case design data.
b Author not affiliated with university
Table . Coding of quality indicators of included single-case design studies.
McCurdy et al. McCurdy et al. Begeny, Krouse,
Begeny & Martens Begeny & Silber Bonfiglio et al. (), (), Ross, and Mitchell Klubnik and Ross and Begeny Begeny, Yeager, & Ross and Begeny
Quality indicator () () () Experiment Experiment () Ardoin () Begeny et al. () () Martínez () ()
Participants Yes (1.0) Yes (1.0) No (0.33) No (0.33) No (0.33) Yes (1.0) No (0.33) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
and setting
Participant Yes (.) Yes (.) No (.) No (.) No (.) Yes (.) No (.) Yes (.) Yes (.) Yes (.) Yes (.)
description
Participant Yes (.) Yes (.) No (.) No (.) No (.) Yes (.) No (.) Yes (.) Yes (.) Yes (.) Yes (.)
selection
Setting Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
description
Dependent Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) No (0.8) No (0.8) Yes (1.0) No (0.8) No (0.8) Yes (1.0) Yes (1.0)
variable
Description Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
Quantifiable Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
Valid and Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
well-described
Measured Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
repeatedly
Interobserver Yes (.) Yes (.) Yes (.) Yes (.) No (.) No (.) Yes (.) No (.) No (.) Yes (.) Yes (.)
agreement
Independent Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
variable
Description Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
Systematic Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
manipulation
Fidelity of imple- Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
mentation
Baseline Yes (1.0) No (0.5) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
Repeated Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
measurement
Description Yes (.) No (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
Experimental Yes (1.0) No (0.66) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
control and
internal
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY
validity
(Continued on next page)
51
52
Table . (Continued)
McCurdy et al. McCurdy et al. Begeny, Krouse,
Begeny & Martens Begeny & Silber Bonfiglio et al. (), (), Ross, and Mitchell Klubnik and Ross and Begeny Begeny, Yeager, & Ross and Begeny
Quality indicator () () () Experiment Experiment () Ardoin () Begeny et al. () () Martínez () ()
J. C. BEGENY ET AL.
Demonstrations of Yes (.) No (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
experimental
effect ( or
more)
Internal validity Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
Pattern of results Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
External Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
validity
Social validity Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0) Yes (1.0)
DV is socially Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
important
Change in DV is Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
socially
important
IV is practical and Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
cost-effective
Used in typical Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.) Yes (.)
contexts
# of indicators 7 5 6 6 5 6 6 6 6 7 7
(absolute
coding)
# of indicators 7.0 6.17 6.33 6.33 6.13 6.8 6.33 6.8 6.8 7.0 7.0
(weighted
coding)
Note. DV = dependent variable; IV = independent variable. Bold text represents a primary quality indicator category.
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 53
or web links to access the protocols and training materials, we felt that sufficed as
detailed documentation to capture the nature of the intervention.
or continued reading difficulties). These findings are not surprising because most
educators recognize the importance of ensuring students are fluent readers by Grade
4 (e.g., Armbruster et al., 2001) and that SG interventions are most often used with
struggling learners prior to a student being considered for special education services
(e.g., Gersten et al., 2008). In terms of identifying students most appropriate to
receive intervention, a variety of screening measures were used, with assessments
of students’ WCPM on seasonal benchmark assessments being most common.
multiple strategies seems like a logical approach based on what is known about
one-on-one reading fluency interventions (Begeny et al., 2010). With this said, a
noticeable finding and gap in this research is that only one study (Begeny, Braun,
et al., 2012) evaluated a structured, manualized intervention program, rather than a
combination of evidence-based strategies and a researcher-developed instructional
protocol. Later we discuss the implications of this finding.
For each SCD study, Table 5 reports the presence or absence of each QI proposed by
Horner et al. (2005). Using the weighted coding scores (with a maximum score of
seven) and an 80% or higher criterion (Sreckovic et al., 2014), each of the SCD stud-
ies can be considered methodologically credible. In fact, all studies had a weighted
coding score above six, and only two studies (Begeny & Silber, 2006; McCurdy,
Daly, Gortmaker, Bonfiglio, & Persampieri, 2007, Experiment 2) had an absolute
coding score below six. Thus, the overall quality of this literature base meets pro-
posed standards. However, there were some studies that did not meet certain stan-
dards. For example, the Begeny and Silber study lacked sufficient information about
the number of baseline data points, and as such it was unclear whether there were
three or more demonstrations of experimental effect. Also, four studies (Bonfiglio,
Daly, Persampieri, & Andersen, 2006; Klubnik & Ardoin, 2010; both experiments in
McCurdy et al., 2007) lacked information about the study participants. Specifically,
for the at-risk participants in those studies, they were nominated by their teachers
as needing reading assistance, but no screening data were reported to more compre-
hensively describe the participant characteristics and selection procedures, thereby
decreasing the ability for the studies to be replicated with precision.
Using the Gersten et al. (2005) criteria, Table 6 reports the QIs of the two GDS
that met our inclusion criteria. Gersten et al. suggested that an acceptable quality
study must meet at least three of the four essential QIs (EQIs) and one of the eight
desirable QIs (DQIs); and a high-quality study must meet at least three EQIs and
four DQIs. Although these criteria can be debated, the suggested guidelines reveal
that neither study in our review received a three or higher with the absolute coding
of the EQIs. However, the Begeny, Braun, et al. (2012) study received a weighted EQI
score of three and a DQI score of seven. Also worth noting, although that study did
not include traditional ES calculations, it did use participant data and national nor-
mative data to systematically examine magnitude of effect and practical significance,
similar to an ES estimate. Although the Kuhn (2005) study lacked several of the QIs,
its peer-reviewed publication may be partially explained by the lack of research on
fluency-based SG interventions at the time of its publication.
Given the QIs from this literature base, there is at least some evidence that com-
bined use of SG instructional components targeting reading fluency can be con-
sidered an EBP. Using the SCD studies, we applied the five criteria suggested by
(Horner et al., 2005) to determine if this body of research can be considered evi-
dence based: “(a) the practice is operationally defined; (b) the context and outcomes
associated with a practice are clearly defined; (c) the practice is implemented with
documented fidelity; (d) the practice is functionally related to change in valued out-
comes; (e) experimental control is demonstrated across a sufficient range of studies,
researchers, and participants to allow confidence in the effect” (p. 176). Aligned with
this last criterion, and consistent with the review by (Sreckovic et al., 2014), we eval-
uated the external validity of this research base (across all studies) by determining
whether the research was conducted by at least three different researchers, with 20 or
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 57
more total participants, and represents at least three geographical regions. With each
of the aforementioned criteria, the SCD studies included in our review demonstrate
that SG instructional practices aimed to improve reading fluency (i.e., the opera-
tionally defined and combined use of LPP, RR, M/R, and PD practices documented
in this literature) can be considered an EBP. Other practices such as oral retell, PF,
and CR also appear to be valuable when combined with LPP, RR, M/R, and PD, but
they have not yet met the external validity criteria specified previously.
For GDS, Gersten et al. (2005) proposed that a practice is evidence based if there
are “at least four acceptable quality studies, or two high quality studies that support
the practice; and there is a 20% confidence interval for the weighted effect size that
is greater than zero” (p. 162). As will be discussed subsequently regarding directions
for future research, there are too few GDS within this literature base to meet these
suggested criteria, and only one GDS study (Begeny, Braun, et al., 2012) arguably
met standards as an acceptable or high-quality study.
As shown in the last column of Table 4, there were five studies that specifically exam-
ined whether comparable SG and one-on-one interventions would result in different
effects (Begeny et al., 2011; Begeny, Yeager, & Martínez, 2012; Klubnik & Ardoin,
2010; Ross & Begeny, 2011, 2015). With each respective study, we first calculated
the number of participants that responded significantly better with either the one-
on-one or SG intervention, compared with the control condition (as shown in the
denominator reported in that column). We then reported the number (numerator)
and percentage of students who, according to the authors, responded as well or bet-
ter during the SG condition compared to the one-on-one condition. Across the five
studies, 79% (19 of 24 students) responded as well or better during the SG inter-
vention. According to the authors of these studies, in the vast majority of cases the
students responded as well to the SG condition as they did to one-on-one (not usu-
ally better), but this is still an important finding. It suggests that a large percentage
of students should fair just as well with the less resource-intensive SG intervention
that targets fluency as they would with a one-on-one intervention.
This should be useful information for practitioners because the vast majority of
research on targeted reading fluency interventions is conducted in a one-on-one
context (Chard et al., 2002; National Institute of Child Health and Human Devel-
opment, 2000; Therrien, 2004), and the studies reviewed in this report should give
useful direction for both researchers and practitioners for developing and further
evaluating SG reading fluency interventions. Also, when considering models of RTI
(Fuchs & Vaughn, 2012; Gersten et al., 2008) some might assume that many students
should respond as well to a less intensive (e.g., Tier 2) type of intervention com-
pared to a more intensive (e.g., Tier 3) intervention. However, as others previously
noted (e.g., Begeny et al., 2011; Ross & Begeny, 2015; Vaughn et al., 2003), differen-
tial responsiveness to interventions of varying intensity is an important empirical
question that cannot be assumed. With regard to reading fluency and other targeted
58 J. C. BEGENY ET AL.
reading interventions, currently very little research has specifically examined the
percentage of students who respond as well to SG intervention as they do to more
intensive forms of intervention. Our study therefore offers additional insight on the
topic by summarizing research that examined the differential effectiveness of com-
parable SG and one-on-one interventions.
sample sizes have been relatively small for a complete body of literature. Also, there
have been no between group comparison studies with sufficient power; in fact, the
one between-group study published (Kuhn, 2005) did not even use inferential statis-
tics to test hypotheses due to such a small sample size. Despite a large proportion of
published school-based intervention studies that do not meet suggested standards
for methodological rigor (e.g., Burns et al., 2012), and despite the extensive chal-
lenges (ethical, resource related, and otherwise) that can come with conducting such
research (Bliss et al., 2008; Villarreal, Gonzalez, McCormick, Simek, & Yoon, 2013),
we hope this current gap in this literature inspires rather than dissuades researchers
from attempting to conduct such studies on SG reading fluency interventions.
Second, the relatively limited populations of students that have been studied in
this research area calls for additional investigations to explore SG fluency inter-
ventions with other populations, including but not limited to upper-grade students
(e.g., Grades 5–10), ELLs living in the United States, students receiving special edu-
cation services, and students of varying ethnicity (including students living outside
of the United States). Third, the existing research only includes one relatively small-
scale study of a manualized, structured intervention program; thus, more studies on
intervention programs are warranted. Although teachers might be able to improve
students’ academic outcomes by reviewing the research and developing their own
protocols based on operational definitions of evidence-based strategies, there
are many reasons related to intervention effectiveness and efficiency that suggest
educators would be best equipped with research-supported intervention programs
versus a protocol of strategies developed by an individual educator, school team, or
district—especially at a Tier 2 level when prescriptive problem solving is arguably
most effective (Begeny, Schulte, & Johnson, 2012). For example, intervention pro-
grams often provide structured protocols, scripted instructions, field-tested reading
passages, forms for recording students’ daily receipt of intervention and academic
progress, embedded assessments, and other related features that ultimately help
to make training more efficient, improve teachers’ implementation integrity, and
ensure that students get explicit, systematic instruction (Begeny, Braun, et al., 2012;
Begeny, Schulte, & Johnson, 2012; Reeves, 2010; Simmons et al., 2011; What Works
Clearinghouse, 2015). Additional research on manualized programs may also help
to enhance operational definitions of specific practices and the determination of
whether those practices are evidence based.
Fourth, to increase the probability of successfully merging research with practice
in this area of reading intervention, future researchers should also ensure there are
more school-based practitioners delivering the interventions, as well as providing
valuable feedback about usability, feasibility, sustainability, and acceptability of the
intervention and the methods of teacher training. Ideally, future researchers would
include school-based practitioners as part of the research team because there are
numerous benefits to this approach to research and it is unfortunately uncommon
in education-based scholarship (Carroll, Skinner, McCleary, Hautau von Mizener,
& Bliss, 2009). For example, of all of the publications included in the present study,
only three authors were listed as being unaffiliated with a university and only two of
60 J. C. BEGENY ET AL.
those individuals (Lynch and Ramsay—both on the same publication) did not have
prior affiliation with one of the lead university researchers.
Overall, our findings suggest that meaningful pathways have been forged in this
area of reading intervention research, and the SCD studies suggest that several SG
instructional practices can be considered evidence-based. However, this body of
empirical literature will undoubtedly benefit from future studies that utilize, for
example, GDS, larger and more diversified samples, evaluations of manualized pro-
grams, and collaborative partnerships with education practitioners.
Note
1. Although we originally set out to complete a meta-analysis of the available research, sev-
eral reasons precluded using meta-analytic procedures. One reason for this decision was
because we realized too few studies used group designs, which would be necessary to facil-
itate the original, traditional, and more common use of meta-analysis (Glass, McGaw, &
Smith, 1981; Rosenthal, 1986). Because 11 of the 13 studies (85%) used an SCD, we then
considered the possibility of including those articles within a meta-analysis for SCDs.
However, a relatively recent special issue within the Journal of School Psychology (April
2014) shed light on this topic and assisted us in coming to the conclusion that a SCD meta-
analysis is not appropriate at this time. The reasons for this decision are as follows. First,
SCD meta-analytic procedures are still emerging, and unlike group-design meta-analytic
procedures, there is no general consensus of a gold standard for SCD meta-analytic pro-
cedures (Kratochwill & Levin, 2014). Second, Maggin and Odom (2014) explained the
necessary requirements to complete a meta-analysis with SCDs and concluded that none
of the proposed analytical methods succeed in meeting all of the criteria; they further indi-
cated that each method requires more development. In addition to the analytical methods
lacking support for their use, they are not readily available to researchers at this time. As
noted by Fisher and Lerman (2014), SCD meta-analytic procedures will require education
researchers to team up with statisticians to correctly use these preliminary and underde-
veloped procedures. Finally, many of those with sophisticated knowledge of meta-analysis
with SCDs (and their limitations) are concerned that the community of SCD scholars will
not accept the proposed meta-analytic procedures as they currently stand (Kratochwill &
Levin, 2014; Shadish, 2014).
References
Armbruster, B., Lehr, F., & Osborne, J. (2001). Put reading first: The research building blocks for
teaching children to read (Kindergarten through Grade 3) (3rd ed.). Washington, DC: National
Institute for Literacy.
Begeny, J. C., Braun, L. M., Lynch, H. L., Ramsay, A. C., & Wendt, J. M. (2012). Initial evidence for
using the HELPS reading fluency program with small instructional groups. School Psychology
Forum: Research in Practice, 6, 50–63.
Begeny, J. C., Hawkins, A. L., Krouse, H. E., & Laugle, K. M. (2011). Altering instruc-
tional delivery options to improve intervention outcomes: Does increased instructional
intensity also increase instructional effectiveness? Psychology in the Schools, 48, 769–785.
https://doi.org/10.1002/pits.20591
Begeny, J. C., Krouse, H. E., Ross, S. G., & Mitchell, R. C. (2009). Increasing elementary-aged
students’ reading fluency with small-group interventions: A comparison of repeated reading,
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 61
listening passage preview, and listening only strategies. Journal of Behavioral Education, 18,
211–228. https://doi.org/10.1007/s10864-009-9090-9
Begeny, J. C., Laugle, K. M., Krouse, H. E., Lynn, A. E., Tayrose, M. P., & Stage, S. A. (2010).
A control-group comparison of two reading fluency programs: The Helping Early Literacy
with Practice Strategies (HELPS) program and the Great Leaps K–2 reading program. School
Psychology Review, 39, 137–155.
Begeny, J. C., & Martens, B. K. (2006). Assisting low-performing readers with a group-based read-
ing fluency intervention. School Psychology Review, 35, 91–107.
Begeny, J. C., Schulte, A. C., & Johnson, K. (2012). Enhancing instructional problem solving: An
efficient system for assisting struggling learners. New York, NY: Guilford Press.
Begeny, J. C., & Silber, J. M. (2006). An examination of group-based treatment packages for
increasing elementary-aged students’ reading fluency. Psychology in the Schools, 43, 183–195.
https://doi.org/10.1002/pits.20138
Begeny, J. C., Yeager, A., & Martínez, R. S. (2012). Effects of small-group and one-on-one reading
fluency interventions with second grade, low-performing Spanish readers. Journal of Behav-
ioral Education, 21, 58–79. https://doi.org/10.1007/s10864-011-9141-x
Berkeley, S., Bender, W. N., Gregg-Peaster, L., & Saunders, L. (2009). Implementation of
response to intervention: A snapshot of progress. Journal of Learning Disabilities, 42, 85–95.
https://doi.org/10.1177/0022219408326214
Bliss, S. L., Skinner, C. H., Hautau, B., & Carroll, E. E. (2008). Articles published in four school
psychology journals from 2000 to 2005: An analysis of experimental/intervention research.
Psychology in the Schools, 45, 483–498. https://doi.org/10.1002/pits.20318
Bonfiglio, C. M., Daly, E. J., Martens, B. K., Lin, L. R., & Corsaut, S. (2004). An experimental analy-
sis of reading interventions: Generalization across instructional strategies, passages, and time.
Journal of Applied Behavioral Analysis, 27, 111–114. https://doi.org/10.1901/jaba.2004.37-111
Bonfiglio, C. M., Daly, E. J., Persampieri, M., & Andersen, M. (2006). An experimental analysis
of the effects of reading interventions in a small group reading instruction context. Journal of
Behavioral Education, 15, 93–109. https://doi.org/10.1007/s10864-006-9009-7
Bramlett, R., Cates, G. L., Savina, E., & Lauinger, B. (2010). Assessing effectiveness and effi-
ciency of academic interventions in school psychology journals. Psychology in the schools, 47,
114–125. https://doi.org/10.1002/pits.20457
Buck, G. H., Polloway, E. A., Smith-Thomas, A., & Cook, K. W. (2003). Prereferral intervention
processes: A survey of state practices. Exceptional Children, 69, 349–360.
Burns, M. K., & Gibbons, K. A. (2008). Implementing response-to-intervention in elementary and
secondary schools. New York, NY: Routledge.
Burns, M. K., Klingbeil, D. A., Ysseldyke, J. E., & Petersen-Brown, S. (2012). Trends in method-
ological rigor in intervention research published in school psychology journals. Psychology
in the Schools, 49, 843–851. https://doi.org/10.1002/pits.21637
Carroll, E. E., Skinner, C. H., McCleary, D. F., Hautau von Mizener, B., & Bliss, S. L.
(2009). Analysis of author affiliation across four school psychology journals from 2000
to 2008: Where is the practitioner research? Psychology in the Schools, 46, 627–635.
https://doi.org/10.1002/pits.20403
Chard, D. J., Vaughn, S., & Tyler, B. J. (2002). A synthesis of research on effective interventions
for building reading fluency with elementary students with learning disabilities. Journal of
Learning Disabilities, 35, 386–406.
Daane, M. C., Campbell, J. R., Grigg, W. S., Goodman, M. J., & Oranje, A. (2005). Fourth-grade
students reading aloud: NAEP 2002 Special Study of Oral Reading. U.S. Department of Edu-
cation, National Center for Education Statistics. Washington, DC: U.S. Government Printing
Office. Retrieved from http://nces.ed.gov/nationsreportcard/pdf/studies/2006469.pdf
62 J. C. BEGENY ET AL.
Denton, C. A. (2012). Response to intervention for reading difficulties in the primary grades:
Some answers and lingering questions. Journal of Learning Disabilities, 45, 232–243.
https://doi.org/10.1177/0022219412442155
Denton, C. A., Cirino, P. T., Barth, A. E., Romain, M., Vaughn, S., Wexler, J., … Fletcher,
J. M. (2011). An experimental study of scheduling and duration of “Tier 2” first-
grade reading intervention. Journal of Research on Educational Effectiveness, 4, 208–230.
https://doi.org/10.1080/19345747.2010.530127
DuFour, R., Eaker, R., Karhanek, G., & DuFour, R. (2004). Whatever it takes: How professional
learning communities respond when kids don’t learn. Bloomington, IN: Solution Tree.
Elbaum, B., Vaughn, S., Tejero Hughes, M., & Watson Moody, S. (2000). How effective are one-
to-one tutoring programs in reading for elementary students at risk for reading failure? A
meta-analysis of the intervention research. Journal of Educational Psychology, 92, 605–619.
https://doi.org/10.1037/0022-0663.92.4.605
Fisher, W. W., & Lerman, D. C. (2014). It has been said that, “There are three degrees of
falsehoods: Lies, damn lies, and statistics”. Journal of School Psychology, 52, 243–248.
https://doi.org/10.1016/j.jsp.2014.01.001
Fletcher, J. M., Stubeing, K. K., Barth, A. E., Denton, C. A., Cirino, P. T., Francis, D. J., & Vaughn,
S. (2011). Cognitive correlates of inadequate response to reading intervention. School Psy-
chology Review, 40, 3–22.
Foorman, B. R., & Torgesen, J. (2001). Critical elements of classroom and small-group instruc-
tion to promote reading success in all children. Learning Disabilities Research & Practice, 16,
203–212. https://doi.org/10.1111/0938-8982.00020
Fuchs, L. S., Fuchs, D., Hosp, M. K., & Jenkins, J. R. (2001). Oral reading fluency as an indicator
of reading competence: A theoretical, empirical and historical analysis. Scientific Studies of
Reading, 5, 239–256.
Fuchs, L. S., & Vaughn, S. (2012). Responsiveness-to-intervention: A decade later. Journal of
Learning Disabilities, 45, 195–203. https://doi.org/10.1177/0022219412442150
Gersten, R., Compton, D., Connor, C. M., Dimino, J., Santoro, L., Linan-Thompson, S., &
Tilly, W. D. (2008). Assisting students struggling with reading: Response to intervention and
multi-tier intervention for reading in the primary grades. A practice guide (NCEE 2009–
4045). Washington, DC: National Center for Education Evaluation and Regional Assis-
tance, Institute of Education Sciences, U.S. Department of Education. Retrieved from
http://ies.ed.gov/ncee/wwc/publications/practiceguides/
Gersten, R., Fuchs, L. S., Compton, D., Coyne, M. S., Greenwood, C., & Innocenti, M. (2005).
Quality indicators for group experimental and quasi-experimental research in special educa-
tion. Exceptional Children, 71, 149–164. https://doi.org/10.1177/001440290507100202
Glass, G. V., McGaw, B., & Smith, M. L. (1981). Meta-analysis in social research. Beverly Hills,
CA: Sage.
Horner, R. H., Carr, E. G., Halle, J., McGee, G., Odom, S., & Wolery, M. (2005). The use of single-
subject research to identify evidence-based practice in special education. Exceptional Chil-
dren, 7, 165–179.
Klubnik, C., & Ardoin, S. P. (2010). Examining immediate and maintenance effects of a reading
intervention package on generalization materials: Individual verses group implementation.
Journal of Behavioral Education, 19, 7–29. https://doi.org/10.1007/s10864-009-9096-3
Kratochwill, T. R., & Levin, J. R. (2014). Meta- and statistical analysis of single-case intervention
research data: Quantitative gifts and a wish list. Journal of School Psychology, 52, 231–235.
https://doi.org/10.1016/j.jsp.2014.01.003
Kuhn, M. R. (2005). A comparative study of small group fluency instruction. Reading Psychology,
26, 127–146. https://doi.org/10.1080/02702710590930492
Kuhn, M. R., & Stahl, S. A. (2003). Fluency: A review of developmental and remedial practices.
Journal of Educational Psychology, 95, 3–21. https://doi.org/10.1037/0022-0663.95.1.3
JOURNAL OF APPLIED SCHOOL PSYCHOLOGY 63
Lonigan, C. J., & Shanahan, T. (2010). Executive summary, Developing early literacy: Report of
the National Early Literacy Panel. Washington, DC: National Institute for Literacy. Retrieved
from https://www.nichd.nih.gov/publications/pubs/documents/NELPSummary.pdf
Maggin, D. M., & Odom, S. L. (2014). Evaluating single-case research data for systematic
review: A commentary for the special issue. Journal of School Psychology, 52, 237–241.
https://doi.org/10.1016/j.jsp.2014.01.002
McCurdy, M., Daly, E., Gortmaker, V., Bonfiglio, C., & Persampieri, M. (2007). Use of brief
instructional trials to identify small group reading strategies: A two experiment study. Journal
of Behavioral Education, 16, 7–26. https://doi.org/10.1007/s10864-006-9021-y
McIntyre, L. L., Gresham, F. M., DiGennaro, F. D., & Reed, D. D. (2007). Treatment
integrity of school-based interventions with children in the Journal of Applied
Behavior Analysis: 1991–2005. Journal of Applied Behavior Analysis, 40, 659–672.
https://doi.org/10.1901/jaba.2007.659-672
Meyer, M. S., & Felton, R. H. (1999). Repeated reading to enhance reading flu-
ency: Old approaches and new directions. Annals of Dyslexia, 49, 283–306.
https://doi.org/10.1007/s11881-999-0027-8
Morgan, P. L., & Sideridis, P. D. (2006). Contrasting the effectiveness of fluency interven-
tions for students with or at risk for learning disabilities: A multilevel random coeffi-
cient modeling meta-analysis. Learning Disabilities Research and Practice, 21, 191–210.
https://doi.org/10.1111/j.1540-5826.2006.00218.x
National Assessment of Educational Progress. (2015). Reading assessment. Retrieved from
http://www.nationsreportcard.gov/reading_math_2015/#reading/acl?grade=4
National Center on Response to Intervention. (2010). Essential components of RTI—A closer
look at response to intervention. Washington, DC: U.S. Department of Education, Office of
Special Education Programs, National Center on Response to Intervention. Retrieved from
http://www.rti4success.org/sites/default/files/rtiessentialcomponents_042710.pdf
National Institute of Child Health and Human Development. (2000). Report of the National Read-
ing Panel: Teaching children to read: An evidence based assessment of the scientific research lit-
erature on reading and its implications for reading instruction: Reports of the subgroups (NIH
Publication No. 00-4754). Washington, DC: U.S. Government Printing Office. Retrieved from
https://www.nichd.nih.gov/publications/pubs/nrp/documents/report.pdf
Reeves, J. (2010). Teacher learning by script. Language Teaching Research, 14, 241–258.
https://doi.org/10.1177/1362168810365252
Rosenthal, R. (1986). Meta-analytic procedures for social research. London: Sage.
Ross, S. G., & Begeny, J. C. (2011). Improving Latino, English language learners’ reading flu-
ency: The effects of small-group and one-on-one intervention. Psychology in the Schools, 48,
604–618. https://doi.org/10.1002/pits.20575
Ross, S. G., & Begeny, J. C. (2015). An examination of treatment intensity with an oral
reading fluency intervention: Do intervention duration and student-teacher instructional
ratios impact intervention effectiveness? Journal of Behavioral Education, 24, 11–32.
https://doi.org/10.1007/s10864-014-9202-z
Sanetti, L. M. H., & DiGennaro, F. M. (2012). Barriers to implementing treatment integrity pro-
cedures in school psychology research: Survey of treatment outcome researchers. Assessment
for Effective Intervention, 37, 195–202. https://doi.org/10.1177/1534508411432466
Sanetti, L. M. H., Gritter, K. L., & Dobey, L. M. (2011). Treatment integrity of interventions with
children in the school psychology literature from 1995 to 2008. School Psychology Review, 40,
72–84.
Seethaler, P. M., & Fuchs, L. S. (2005). A drop in the bucket: Randomized controlled trials test-
ing reading and math interventions. Learning Disabilities Research and Practice, 20, 98–102.
https://doi.org/10.1111/j.1540-5826.2005.00125.x
64 J. C. BEGENY ET AL.