Professional Documents
Culture Documents
DOI: 10.1002/pits.22100
RESEARCH ARTICLE
York Abstract
2 Oklahoma State University The purpose of the current study is to determine whether single-
Correspondence case intervention studies targeting reading fluency, ranked by tra-
Benjamin Solomon, PhD, University at Albany, ditional outcome metrics (i.e., effect sizes derived from phase dif-
Division of School Psychology, Department of ferences), were discrepant with rankings based on instructional effi-
Educational and Counseling Psychology, State
University of New York Education 232, 1400
ciency, including growth per session and minutes of instruction. Con-
Washington Ave. Albany, NY 12222. verging with prior findings, we found great variability in reported
Email: bgsolomon@albany.edu sessions and minutes of instruction across studies, as well as diver-
gences in rankings based on outcome variables. These findings raise
questions as to how literature syntheses on the topic of academic
intervention are interpreted and how selection of evidence-based
intervention occurs.
KEYWORDS
cumulative instructional time, effect sizes, oral reading fluency
A fundamental responsibility of trainers and school-based practitioners is the promotion and dissemination of
evidence-based intervention (EBI), which is in line with a national call for use of educational practices with a prepon-
derance of high-quality evidence (e.g., Individuals with Disabilities in Education Act, 2004). Such a focus has spurred
increased emphasis on the use of experimental designs to generate valid causal conclusions regarding treatment effec-
tiveness (Kratochwill et al., 2010; Shadish, Cook, & Campbell, 2001). However, the demand for a higher and consistent
standard regarding the identification of EBI has also put focus on how study findings are summarized and presented to
consumers. That is, academic growth under the presence of intervention can be summarized in different ways, which
can be misleading if nuisance variables are not controlled and comparative metrics equated over studies (Bramlett,
Cates, Savina, & Lauinger, 2010; Poncy et al., 2015; Skinner, Belfiore, & Watson, 1995; Wolery, Busick, Reichow, & Bar-
ton, 2010). Such confusion can lead to the selection of EBIs that are less than optimal given the interventions available
in the peer-reviewed literature.
Skinner et al. (1995) identified CIT, or the amount of time required for an intervention to be administered as described,
as an overlooked variable in the comparative analysis of EBIs. One intervention may appear more effective than
another partially or totally because its reported dosage, in terms of required session length and number of sessions,
is greater in one study than in that of a comparison intervention, raising the question as to what would happen had
CIT been equated. Practitioners may then elect to use interventions that have been shown to be effective, but increase
learning rates to an inferior level relative to another available intervention. To remedy this, the authors encouraged
shifting discussion of outcomes from phase differences or pre- to post-improvement to rates of learning during inter-
vention, i.e., the difference in skill performance per minute of instruction from one instructional environment, such as
the baseline, to the next, such as the intervention.
Consideration of CIT is most relevant in comparative research where two or more interventions are contrasted.
Classically, when an intervention is compared to a no-treatment group or baseline phase, CIT is not a concern because
the “business-as-usual” condition represents no time away from everyday instruction (Poncy et al., 2015). In this case,
although the CIT of the intervention condition is of interest, the rate of learning cannot be compared to the no-
treatment condition, where no time was spent in intervention. However, when two interventions are compared within
or across experiments, comparisons of CIT become meaningful, as both interventions represent time away from the
student's primary learning environment and the interventions can be precisely timed. Accompanying traditional out-
come metrics with such rate-based metrics allows for equitable comparisons across studies (Skinner, 2008; Skinner
et al., 1995; Skinner, Fletcher, & Henington, 1996). This is particularly important given that instructional time in schools
is at a premium (Fisher et al., 1978; Gettinger & Ball, 2008). A basic tenet of the EBI movement is that research should
increasingly move toward establishing the best intervention for a given instructional deficit. Therefore, comparative
studies that contrast multiple interventions should become more frequent as the technology of EBI moves forward
and discussion of rates of learning become more relevant.
Supporting this conceptual argument is a line of applied studies that have targeted CIT as the time function by which
change in the dependent variable (DV) was measured. For example, Nist and Joseph (2008) specified growth along both
the x-axis of intervention sessions and CIT, comparing improvement in words mastered under the presence of flashcard
drill and practice and incremental rehearsal interventions. Under the context of traditional session-by-session growth,
incremental rehearsal was most effective. However, in the latter analysis, flashcard drill and practice was most efficient.
The authors theorized that this was because the use of known words in incremental rehearsal would draw out CIT as
these words do not represent learning opportunities.
Additional studies comparing growth by way of sessions and CIT appeared around this time (Cates et al., 2003;
Joseph & Nist, 2006; Volpe, Mulé, Briesch, Joseph, & Burns, 2011; Yaw et al., 2014). The message across these stud-
ies was fairly consistent: rate-based metrics need to be considered and often diverge from the outcomes commonly
reported, such as phase differences or rate of growth per session, not only when experimentally comparing interven-
tions but also within the broader context of EBI selection, literature review, and meta-analysis.
The implications of CIT are particularly relevant in the context of meta-analysis. Meta-analysis is a framework for
pooling and comparing treatment effects across studies by standardizing the between groups or phase difference
within studies (i.e., the effect size [ES]), commonly between a no-treatment and treatment condition. Advances in
the technology by which single-case (SC) studies can be summarized using ESs have resulted in a sharp increase in
such studies. An ES is typically selected that reflects the difference between an adjacent intervention phase and the
baseline phase, mirroring their group design counterparts (i.e., Cohen's d, Hedges’ g), and a variety of parametric and
nonparametric techniques are used to accomplish this task (e.g., Parker, Vannest, Davis, & Sauber, 2011; Shadish,
SOLOMON ET AL . 153
Rindskopf, & Hedges, 2008). However, an assumption of comparative meta-analysis is that included studies are
sufficiently exchangeable, meaning that included ESs of the sample are random draws from the possible population
of all studies. Differences among studies may occur, but are not predictable prior to analysis (Cameron et al., 2015;
Higgens, Thompson, & Spiegelhalter, 2009). Large discrepancies in CIT likely spur a violation of this assumption.
Several meta-analyses of reading intervention studies using group designs have acknowledged the potential influ-
ence of CIT by conducting moderator analysis, typically categorizing intervention studies into groups based on arbi-
trary cut-offs for the number of sessions or hours of interventions reported (e.g., Flynn, Zheng, & Swanson, 2012;
Scammacca, Roberts, Vaughn, & Stuebing, 2015; Suggate, 2016). These studies are discrepant in their findings, with
some reporting the duration of intervention-moderated effects and others reporting that the duration of intervention
was not influential. For example, Berkeley, Scruggs, and Mastropieri (2010) reported that medium-length interventions
(1 week to 1 month) were more effective on average than short (<1 week) or long interventions (>1 month). However,
this traditional approach does not equalize the effect of CIT on other, more meaningful comparisons (i.e., comparative
effectiveness of different interventions). This is also in contrast to how Skinner et al. (1995) envisioned evaluating the
comparative quality of interventions: as operating on rates of growth based on time spent in intervention and fixing
dosage across all other dimensions of analysis. These estimates, pooled by intervention or a relevant moderator, may
obscure the instructional efficiency of individual interventions.
To demonstrate how CIT potentially serves as a nuisance variable in research synthesis, Poncy et al. (2015) reviewed
SC intervention studies targeting math fluency, including eight different types of interventions. These authors ranked
the associated outcome metrics by traditional ES (i.e., phase comparison) and when the rate of growth was measured
as ES/CIT. Their ranking of interventions from most to least effective varied substantially based on the metric used. The
authors concluded that meta-analysis and applied comparative study to increase math fluency is not straightforward;
in either case, learning rates should be considered. Poncy et al. (2015), however, had two limitations warranting further
research. First, these authors focused on the subset of the peer-reviewed literature that described math interventions
and associated digits correct per minute outcomes. It is possible that findings in literacy-based interventions are dif-
ferent, whereby traditional metrics of growth (the ESs) mirror rate-based metrics, and thus CIT is not a practical cause
for concern.
Second, Poncy et al. (2015) defined their rate metric as summative ESs divided by CIT. This is in line with most extant
SC meta-analyses, which generally have used more traditional ESs that reflect average phase differences or, more com-
monly, phase nonoverlap. However, doing so has limitations. These rates have no identifiable variance and can nei-
ther be tested for significance nor pooled using traditional meta-analytic procedures. They are primarily descriptive in
nature. Fortunately, these hurdles can be overcome by, alternatively, using simple growth models, such as those out-
lined by Moeyaert et al. (2015), who provided adjustments to traditional regression analysis to capture such effects
specifically for SC designs. Rather than dividing the calculated ES by CIT post hoc, a time-series regressor is specified
in an ordinary least squares regression (OLS) that mirrors the rate metric in question (e.g., CIT accumulated over the
course of the intervention [see Method section]). Unlike dividing ES by CIT, results of this method are easy to visual-
ize, result in standard OLS beta weights, and such effects can be pooled using traditional meta-analytic procedures.
As an example, Figure 1 yields growth estimates for subject two from Walcott, Marett, and Hessel (2014), which is
an included article from the current study. Here, we demonstrate how regressions isolating (1) average growth across
phases, (2) growth per session of intervention, or (3) growth per 15 minutes of intervention results in different esti-
mates of effect and, therefore, different interpretations of the outcome data.
4 CURRENT STUDY
Isolating the instructional efficiency of an intervention has direct implications for researchers and practitioners engag-
ing in the testing and selection of EBI. The current study builds upon Poncy et al. (2015) by exploring CIT atten-
uation across the body of reading fluency intervention work for traditional growth and rates of learning (CIT) and
154 SOLOMON ET AL .
F I G U R E 1 Linear regression methods by which to capture study effects. The upper panel demonstrates change mea-
sured as a phase difference. The lower right panel shows change captured for session-by-session growth. The bottom
left panel envisions growth by means of cumulative instructional time, specified as per 15 minutes of instruction
session-by-session growth across SC studies. The purpose of the current study is to investigate the extant literature
for reading fluency interventions, isolating and comparing various rates of learning using growth models. Our research
questions were:
1. Do studies ranked by traditional outcome metrics (i.e., phase differences) differ in order when ranked as rate metrics
(learning rate defined session-by-session or by CIT)?
2. What is the statistical relationship among these outcome metrics?
5 METHOD
1. The article describes an SC experimental reading intervention. Experimental is defined as the study describing the
evaluation of at least two conditions entailing at least three demonstrations of behavioral change at separate time
points (Kratochwill et al., 2010). Intervention is defined as any modification of literacy instruction beyond described
typical classroom instruction for a subset of students.
2. A baseline condition is included that entailed no treatment. All studies that met this qualification also included for-
mative ORF as a DV, which otherwise would have served as an additional rule-in criteria.
3. The duration of the intervention session is reported in minutes or seconds as well as the number of sessions.
Identified articles from the Boolean searches were first screened for appropriateness based on their title and
abstracts. Remaining articles were then sequentially screened by rule-in criteria as described previously. Figure 2
details this process and associated reliability (kappa) from an independent rater for each criterion. An additional three
SC articles were removed upon further inspection because ORF progress monitoring occurred on trained probes (e.g.,
final read of repeated readings), whereas all other included studies examined generalization probes. The final pool of
articles consisted of 18 SC studies. These studies were then organized into homogenous groups (see Table 1). An inde-
pendent researcher reviewed these categories and then grouped studies blind to the coding of the first author. Agree-
ment between the raters was 100%.
F I G U R E 2 Article search and elimination process. Numbers in parentheses represent the number of articles elimi-
nated at each step
156 SOLOMON ET AL .
Intervention Description
Repeated reading (RR) group A target probe is presented to a student who reads it multiple times with
some level of corrective feedback. Some minor variations in the primary
procedure exist.
RR+ Additional components are added that serve different purposes and may
significantly increase cumulative instructional time. Examples include
listening passage preview or phrase drill error correction.
Peer administered Interventions included peer-assisted learning strategies (Fuchs et al.,
2001) that use explicit instruction and repetition of passages that are
administered by a peer student.
Direct instruction Broad-based explicit instruction with time devoted to teaching multiple
evidence-based pillars of reading. For example, phonics (15 minutes),
fluency (10 minutes), and vocabulary (5 minutes).
Earobics Interactive computer-based instruction that focuses on multiple skills
with a general focus on phonics.
Folding in flashcarding A flashcard-based acquisition intervention entailing presentations of
known and unknown words. Self-graphing was also used as reported
presently.
Read naturally software edition RRs with additional comprehension question and vocabulary preview.
Because the intervention is computer administered, it was separated
from the RR studies.
Video modeling Students view themselves reading prior to practicing fluency passages in
conjunction with performance feedback and goal setting.
6 PROCEDURES
example, if an intervention was reported to take 20 minutes per session across four sessions, the new time predictor
increased accordingly: 20/15, 40/15, 60/15, 80/15, or 1.33, 2.67, 4.00, 5.33. This model is an effective mathematical
representation of Skinner et al.’s 1995 difference in the rate of learning. Target metrics included bphase , gphase, bsession ,
gsession , bCIT , and gCIT . Each initial adjacent comparison from baseline to intervention was included. All SC studies were
multiple baseline designs. If reported instructional session time varied by subject or session, the mean value of the
reported range was used.
Eight different intervention types were identified. We aggregated results into tables and figures as described
later. Next, and most importantly, we calculated correlations across outcome DVs, correcting for family-wise error,
so as to explore covariance between growth per minute of intervention, growth per session, and the traditional phase
difference ES. In this analysis, each study participant was treated as an individual subject. A related secondary analysis
investigated whether the additional predictors (i.e., the slope and interaction term) explained additional variance in
ORF scores beyond that of a linear model with only the phase predictor. Therefore, the current study analyzed results
at two levels: by intervention type and by subject, collapsed across interventions.
7 RESULTS
Note: RR-G, group-administered repeated readings; PEER, peer-administered interventions; RR+, individual repeated readings
plus additional components; CR, corrective reading; FI, folding-in; VM, video modeling; RNSE, read naturally software edition;
DI, comprehensive direct instruction; PF, performance feedback.
Note: CIT, cumulative instructional time; CI, 95% confidence interval; RR-G, group-administered repeated readings; PEER,
peer-administered interventions; RR+, individual repeated readings plus additional components; CR, corrective reading; FI,
folding-in; VM, video modeling; RNSE, read naturally software edition; DI, comprehensive direct instruction; Numbers in
parentheses represent confidence intervals.
SOLOMON ET AL . 159
Note: RR-G, group-administered repeated readings; PEER, peer-administered interventions; RR+, individual repeated read-
ings plus additional components; CR, corrective reading; FI, folding-in; VM, video modeling; RNSE, read naturally software edi-
tion; DI, comprehensive direct instruction; UST, unstandardized effect; ST, standardized effect. Interventions ranked lowest to
highest.
Note: UST, unstandardized effect; ST, standardized effect. Family-wise error controlled with the Holm's procedure.
a
Significant at 0.05 level.
b
Significant at 0.01 level.
model. This suggests that the full model was, on average, a better solution to explaining the data, even when corrected
for the additional predictors. Results based on CIT were identical.
8 DISCUSSION
The purpose of the current study is to replicate and extend the findings of Poncy et al. (2015) by investigating and com-
paring instructional rates across reading fluency interventions by both sessions and minutes of reported intervention
exposure using growth models. Poncy et al. (2015) found that estimates of effect evaluated by means of traditional
ESs were discrepant with rates of learning, measured as ES/CIT. These authors therefore cautioned that the varied lev-
els of CIT required to administer academic interventions may obscure rates of learning when engaging in comparative
intervention research and EBI selection. Similar results were found in the present study. We caution, as did Poncy et al.
(2015), that the purpose of the current study was to evaluate metrics of growth, not the relative quality of specific
interventions. Intervention effects were aggregated to demonstrate variability across metrics.
from one of the highest to one of the lowest ranked interventions based on the type of growth examined for the stan-
dardized effect. At the same time, video modeling increased in rank when CIT was taken into account across both rate
metrics and unstandardized and standardized effects. This particular example serves as an excellent demonstration of
why acknowledging CIT is important. Given this intervention's brief CIT (eight sessions at 4 minutes each), the rate
of learning, 0.71 WCPM/15 minutes, was well beyond that of many other interventions even though, by nature of
the experimental design, its phase difference was average for the pool of studies. Simply acknowledging the number
of sessions—information easily acquired from any SC study—would better approximate learning rate in this case. In
the context of traditional ESs used in meta-analysis, this subtlety would likely be overlooked. The analysis conducted
presently acknowledges this impressive effect given the brevity of the intervention.
Correlational analysis also supports these conclusions. Although growth per session and growth per minute corre-
lated at a significant level, the relationships were far from perfect (i.e., 𝜌 = 1.00), which is in line with the findings of Nist
and Joseph (2008) and other similar studies, which found practical variation in growth rates based on whether growth
over sessions or CIT were examined. Of concern, the relationship between phase differences and growth by session
or per 15 minutes of instructional time was practically nonexistent. The one exception was the significant relationship
between the unstandardized phase and phase × session effect. However, even this correlation, 𝜌 = 0.33, was practically
low. If rate of growth per instructional minute is an ideal target of analysis (Skinner et al., 1995), the current analysis
demonstrated that traditional outcome metrics do not reflect that growth. Consumers of EBI should not make infer-
ences regarding intervention efficiency based on descriptions of phase differences. In the context of individual studies
or meta-analysis, it is important that rate-based metrics accompany more traditional analyses and that both be dis-
cussed. Accounting for session-by-session growth is a more precise estimate, quite achievable using the growth mod-
els described presently, and increasingly common (Moeyaert, Ferron, Beretvas, & Van den Noortgate, 2014; Moeyaert
et al., 2015; Shadish, Kyse, & Rindskopf, 2013); however, this metric does not fully overlap with CIT.
8.3 Limitations
The current study must be considered in light of a number of limitations. For one, although each article included
reported session time to the minute, there is undoubtedly some errors in these reports. Given the importance of CIT
and the findings of the current study, we encourage researchers to report CIT and rate-based metrics with greater fre-
quency. Similarly, we reiterate that the current study's focus was on the nature of various outcome variables, not the
comparative quality of EBIs. Some interventions were represented by relatively few studies. We caution that readers
should not draw conclusions regarding EBI from these rankings.
Rate-based outcome metrics may be overly conservative in the event that a researcher purposefully extends an
intervention phase to solidify their demonstration of operant control of reading fluency behavior. A treatment effect
may asymptote, thereby resulting in a reduction in the rate of learning for every additional intervention session that
occurs. Relatedly, the use of rate-based metrics carries a strong assumption of linearity, that is, improvement in fluency
occurs identically for each additional minute or session of dosage. We therefore stress that rate-based metrics and
summaries of CIT accompany traditional estimates of effect. When this information is discrepant, it offers researchers
an opportunity to explain why this is. Future research could also experiment with calculating CIT from the beginning
of the intervention phase to the point at which the interventionist identifies the student as achieving the instruc-
tional goal, with further observations explicitly identified as a demonstration of effect (Poncy et al., 2015). There is
some precedent for this but far more research into such analytic models is required (see Sullivan, Shadish, & Steiner,
2015).
Unlike prior meta-analyses of literacy interventions, we purposefully chose not to include any additional modera-
tors of effect. Given the multitude of interventions examined and moderators analyzed in prior studies with paradoxi-
cal findings, it is unlikely that the influence of the moderators could be parsed out from the effect of the interventions
themselves, nor where such moderators are of interest in addressing the research questions. Nonetheless, moderator
analysis may prove of value in future comparative studies examining rate-based metrics.
162 SOLOMON ET AL .
9 CONCLUSION
The purpose of the current study is to examine how traditional measures of effect correspond with the rate-based
metrics advocated for by Skinner and colleagues across different research designs. Converging with the findings of
Poncy et al. (2015), we found substantial shifts in rankings based on outcome metrics, in addition to wide ranges in CIT
across articles. We therefore advise those conducting comparative intervention synthesis to explicitly acknowledge
CIT either through parallel analysis with traditional estimators, which is a different approach than that of simple mod-
erator analysis, or as a primary outcome variable. This is particularly relevant for studies focusing on skill acquisition;
slope, not level, differences are expected. The current study also demonstrates the flexibility of using growth models,
of which there is an increasingly wide assortment, to model SC intervention effects within or across studies.
Finally, we recommend practitioners locate information on the required CIT of interventions and, when possible,
compare rates of learning under the presence of intervention in addition to or instead of simple phase differences. This
is not to say that interventions that take less time to administer are inherently better. Furthermore, there are additional
factors associated with intervention success that practitioners may want to consider in selecting EBI, such as long-term
retention (Nist & Joseph, 2008; Zaslofsky, Scholin, Burns, & Varma, 2016) or ease of administration. Nonetheless, we
recommend that those comparing EBI treat rate of learning as a primary outcome variable, just as phase differences
have historically been reported and interpreted. Considering such information will result in the selection of interven-
tions that are most efficient in resolving instructional deficits, resulting in the optimal use of instructional resources in
the school.
ORCID
Benjamin G. Solomon http://orcid.org/0000-0002-8457-1112
REFERENCES
*Studies included in the analysis.
*Albers, C. A., & Hoffman, A. (2012). Using flashcard drill methods and self-graphing procedures to improve the reading perfor-
mance of English language learners. Journal of Applied School Psychology, 28, 367–388.
*Allen-DeBoer, R. A., Malmgren, K. W., & Glass, M. (2006). Reading instruction for youth with emotional and behavioral disor-
ders in a juvenile correctional facility. Behavioral Disorders, 32, 18–28.
*Barton-Ardwood, S. M., Wehby, J. H., & Falk, K. B. (2005). Reading instruction for elementary-age students with emotional and
behavioral disorders: Academic and behavioral outcomes. Exceptional Children, 72, 7–27.
*Bennett, J. G., Gardner, R., III, Cartledge, G., Ramnath, R., & Council, M. R., III. (2017). Second-grade urban learners: Preliminary
findings for a computer-assisted, culturally relevant, repeated reading intervention. Education and Treatment of Children, 40,
145–186.
Berkeley, S., Scruggs, T. E., & Mastropieri, M. A. (2010). Reading comprehension instruction for students with learning disabili-
ties, 1995–2006: A meta-analysis. Remedial and Special Education, 31, 423–436.
Bramlett, R., Cates, G. L., Savina, E., & Lauinger, B. (2010). Assessing effectiveness and efficiency of academic interventions in
school psychology journals: 1995–2005. Psychology in the Schools, 47, 114–125.
*Bray, M. A., Kehle, T. J., Spackman, V. S., & Hintze, J. M. (1998). An intervention program to increase reading fluency. Special
Services in the Schools, 14, 105–125.
Cameron, C., Fireman, B., Hutton, B., Clifford, T., Coyle, D., Wells, G., … Toh, S. (2015). Network meta-analysis incorporating
randomized controlled trials and non-randomized comparative cohort studies for assessing the safety and effectiveness of
medical treatments: Challenges and opportunities. Systemic Reviews, 4, 147–155.
Cates, G. L., Skinner, C. H., Steuart Watson, T., Meadows, T. J., Weaver, A., & Jackson, B. (2003). Instructional effectiveness and
instructional efficiency considerations for data- based decision making: An evaluation of interspersing procedures. School
Psychology Review, 23, 601–616.
*De la Colina, M. G., Parker, R. I., Hasbrouck, J. E., & Lara-Alecio, R. (2001). Intensive intervention in reading fluency for at-risk
beginning Spanish readers. Bilingual Research Journal, 25, 503–538.
SOLOMON ET AL . 163
Fisher, C., Filby, N., Marliave, R., Cahen, L., Dishaw, M., Moore, J., & Berliner D. (1978). Teaching behaviors, academic learning
time, and student achievement: Final report of phase III-B of the Beginning Teacher Evaluation Study. San Francisco, CA: Far West
Laboratory for Educational Research and Development.
Flower, A., McKenna, W. J., & Upreti, G. (2016). Validity and reliability of GraphClick and DataThief III for data extraction.
Behavior Modification, 40, 396–413.
Flynn, L. J., Zheng, X., & Swanson, H. L. (2012). Instructing struggling older readers: A selective meta-analysis of intervention
research. Learning Disabilities Research & Practice, 27, 21–32.
Fuchs, D., Fuchs, L. S., Thompson, A., Svenson, E., Yen, L., & Al Otaiba, S., … Saenz, L. (2001). Peer-assisted learning strategies in
reading: extensions for first-grade, kindergarten, and high school. Remedial and Special Education, 22, 15–21.
Gettinger, M., & Ball, C. (2008). Best practices in increasing academic engaged time. In A. Thomas, & J. Grimes (Eds.), Best prac-
tices in school psychology V (pp. 1043–1058). Bethesda, MD: The National Association of School Psychologists.
GraphClick [Computer software]. (2012). Retrieved from http://www.arizona-software.ch/graphclick
*Harris, P. J., Oakes, W. P., Lane, K. L., & Rutherford, R. B., Jr. (2009). Improving the early literacy skills of students at risk for
internalizing or externalizing behaviors with limited reading skills. Behavioral Disorders, 34, 72–90.
Hedges, L. V. (1981). Distribution theory for Glass's estimator of effect size and related estimators. Journal of Educational Statis-
tics, 6, 107–128.
Higgens, J. P. T., Thompson, S. G., & Spiegelhalter, D. J. (2009). A re-evaluation of random- effects meta-analysis. Journal of the
Royal Statistical Society. Series A, 172, 137–159.
*Hofstadter-Duke, K. L., & Daly, E. J., III. (2011). Improving oral reading fluency with a peer- mediated intervention. Journal of
Applied Behavior Analysis, 44, 641–646.
*Hua, Y., Hendrickson, J. M., & Therrien, W. (2012). Effects of combined reading and question generation on reading fluency
and comprehension of three young adults with autism and intellectual disability. Focus on Autism and Other Developmental
Disabilities, 27, 135–146.
Hunter, J. E., & Schmidt, F. L. (2004). Methods of meta-analysis: Correcting error and bias in research synthesis (2nd ed.). Thousand
Oaks, CA: Sage.
Individuals with Disabilities Education Act, 20 U.S.C. § 1400 (2004).
Joseph, L. M., & Nist, L. M. (2006). Comparing the effects of unknown-known ratios on word reading learning versus learning
rates. Journal of Behavioral Education, 15, 69–79.
*Keyes, S. E., Cartledge, G., Gibson, L., Jr., & Robinson-Ervin, P. (2016). Programming for generalization of oral reading fluency
using computer-assisted instruction and changing fluency criteria. Education and Treatment of Children, 39, 141–172.
*Klingbeil, D. A., Moeyaert, M., Archer, C. Y., Chimboza, T. M., & Zwolski, S. A., Jr. (2017). Efficacy of peer-mediated incremental
rehearsal for English language learners. School Psychology Review, 46, 122–140.
Kratochwill, T. R., Hitchcock, J., Horner, R. H., Levin, J. R., Odom, S. L., Rindskopf, D. M & Shadish, W. R. (2010). Single-case
designs technical documentation. Retrieved from http://ies.ed.gov/ncee/wwc/pdf/wwc_scd.pdf
*Lo, Y., Cooke, N. L., & Starling, A. L. P. (2011). Using a repeated reading program to improve generalization of oral reading
fluency. Education and Treatment of Children, 34, 115–140.
Moeyaert, M., Ferron, J. M., Beretvas, N., & Van den Noortgate, W. (2014). From a single- level analysis to a multilevel analysis
of single-case experimental designs. Journal of School Psychology, 52, 191–211.
Moeyaert, M., Ugille, M., Ferron, J. M., Onghena, P., Heyvaert, M., Beretvas, S. N., & Van den Noortgate, W. (2015). Estimating
intervention effects across different types of single-subject experimental designs: Empirical illustration. School Psychology
Quarterly, 30, 50–63.
*Musti-Rao, S., Hawkins, R. O., & Barkley, E. A. (2009). Effects of repeated readings on the oral reading fluency of urban fourth-
grade students: Implications for practice. Preventing School Failure, 54, 12–23.
*Neddenriep, C. E., Fritz, A. M., & Carrier, M. E. (2010). Assessing for generalized improvements in reading comprehension by
intervening to improve reading fluency. Psychology in the Schools, 48, 14–27.
Nist, L., & Joseph, L. M. (2008). Effectiveness and efficiency of flashcard drill instructional methods on urban first-graders’ word
recognition, acquisition, maintenance, and generalization. School Psychology Review, 37, 294–308.
No Child Left Behind Act of 2001, 20 U.S.C. § 6319 (2002).
*Oddo, M., Barnett, D. W., Hawkins, R. O., & Musti-Rao, S. (2010). Reciprocal peer tutoring and repeated reading: Increasing
practicality using student groups. Psychology in the Schools, 47, 842–858.
*O'Shaughnessy, T. E., & Swanson, H. L. (2000). A comparison of two reading interventions for children with reading disabilities.
Journal of Learning Disabilities, 33, 257–277.
164 SOLOMON ET AL .
Parker, R. I., Vannest, K. J., Davis, J. L., & Sauber, S. B. (2011). Combining nonoverlap and trend for single-case research: Tau-U.
Behavior Therapy, 42, 284–299.
Poncy, B. C., Solomon, B., Duhon, G., Skinner, C., Moore, K., & Simons, S. (2015). An analysis of learning rate and curricular scope:
Caution when choosing academic interventions based on aggregated outcomes. School Psychology Review, 44, 289–305.
Rakap, S., Rakap, S., Evran, D., & Cig, O. (2016). Comparative evaluation of the reliability and validity of three data extraction
programs: UnGraph, GraphClick, and Digitize It. Computer in Human Behavior, 55, 159–166.
*Ross, S. G., & Begeny, J. C. (2015). An examination of treatment intensity with an oral reading fluency intervention: Do inter-
vention duration and student–teacher instructional ratios impact intervention effectiveness. Journal of Behavioral Educa-
tion, 24, 11–32.
Scammacca, N. K., Roberts, G., Vaughn, S., & Stuebing, K. K. (2015). A meta-analysis of interventions for struggling readers in
grades 4–12: 1980–2011. Journal of Learning Disabilities, 48, 369–390.
Shadish, W. R., Cook, T. D., & Campbell, D. T. (2001). Experimental and quasi-experimental designs for generalized causal inference.
Belmont, CA: Wadsworth.
Shadish, W. R., Kyse, E. N., & Rindskopf, D. M. (2013). Analyzing data from single-case designs using multilevel models: New
applications and some agenda items for future research. Psychological Methods, 18, 385–405.
Shadish, W. R., Rindskopf, D. M., & Hedges, L. V. (2008). The state of the science in the meta-analysis of single-case experimental
designs. Evidence-Based Communication Assessment and Intervention, 2, 188–196.
Skinner, C. H. (2008). Theoretical and applied implications of precisely measuring learning rates. School Psychology Review, 37,
309–314.
Skinner, C. H., Belfiore, P. J., & Watson, T. S. (1995). Assessing the relative effects of interventions in students with mild disabil-
ities: Assessing instructional time. Journal of Psychoeducational Assessment, 20, 346–357.
Skinner, C. H., Fletcher, P. A., & Henington, C. (1996). Increasing learning rates by increasing student response rates: A summary
of research. School Psychology Quarterly, 11, 313–325.
Suggate, S. P. (2016). A meta-analysis of the long-term effects of phonemic awareness, phonics, fluency, and reading compre-
hension interventions. Journal of Learning Disabilities, 49, 77–96.
Sullivan, K. J., Shadish, W. R., & Steiner, P. M. (2015). An introduction to modeling longitudinal data with generalized additive
models: Applications to single-case designs. Psychological Methods, 20, 26–42.
Van den Noortgate, W., & Onghena, P. (2008). A multilevel meta-analysis of single-subject experimental design studies. Evidence
Based Communication Assessment and Intervention, 2, 142–151.
Volpe, R. J., Mulé, C. M., Briesch, A. M., Joseph, L. M., & Burns, M. K. (2011). A comparison of two flashcard drill methods target-
ing word recognition. Journal of Behavioral Education, 20, 117–137.
*Walcott, C. M., Marett, K., & Hessel, A. B. (2014). Effectiveness of a computer-assisted intervention for young children with
attention and reading problems. Journal of Applied School Psychology, 30, 83–106.
Wolery, M., Busick, M., Reichow, B., & Barton, E. E. (2010). Comparison of overlap methods for quantitatively synthesizing
single-subject data. The Journal of Special Education, 44, 18–28.
Yaw, J., Skinner, C. H., Maurer, K., Skinner, A. L., Cihak, D., Wilhoit, B., … Booher, J. (2014). Measurement scale influences in the
evaluation of sight word reading interventions. Journal of Applied Behavior Analysis, 47, 360–379.
Zaslofsky, A. F., Scholin, S. E., Burns, M. K., & Varma, S. (2016). Comparison of opportunities to respond and generation effect as
potential causal mechanisms for incremental rehearsal with multiplication combinations. Journal of School Psychology, 55,
71–78.
How to cite this article: Solomon BG, Poncy BC, Caravello D, Schweiger EM. Examining Learning Rates
in the Evaluation of Academic Interventions that Target Reading Fluency. Psychol Schs. 2018;55:151–164.
https://doi.org/10.1002/pits.22100