What this is
- This review evaluates targeted school-based interventions aimed at improving reading and mathematics for students in Grades K-6 who face academic difficulties.
- It synthesizes findings from 607 studies, focusing on the effectiveness of various instructional methods.
- The review identifies and as the most effective strategies for enhancing academic outcomes.
Essence
- Targeted school-based interventions for students with or at risk of academic difficulties in Grades K-6 yield positive effects on reading and mathematics. Peer-assisted and are particularly effective methods.
Key takeaways
- Interventions targeting at-risk students in Grades K-6 show statistically significant improvements in academic performance. The average for short-term outcomes is 0.30, indicating meaningful gains.
- and consistently demonstrate the largest effect sizes, ranging from 0.35 to 0.45, making them effective components of academic interventions.
- Follow-up effects remain positive, with an average of 0.27, suggesting that the benefits of these interventions can persist beyond immediate post-intervention assessments.
Caveats
- Many studies included in the review exhibited a moderate to high risk of bias, which may affect the reliability of the reported effect sizes.
- Most follow-up outcomes pertain to and reading measures, limiting the generalizability of findings to other instructional methods or subjects.
- The review primarily includes studies from the United States, which may not fully represent the effectiveness of interventions in different educational contexts.
Definitions
- Effect Size (ES): A quantitative measure of the magnitude of a phenomenon. In this review, it indicates the average difference in academic performance between intervention and control groups.
- Peer-assisted instruction: An instructional approach where students work together, often in pairs, to support each other's learning.
- Small-group instruction: Teaching method where a small number of students receive instruction together, allowing for more personalized attention.
AI simplified
PLAIN LANGUAGE SUMMARY
Targeted school‐based interventions improve achievement in reading and maths for at‐risk students in Grades K‐6
School‐based interventions that target students with, or at risk of, academic difficulties in kindergarten to Grade 6 have positive effects on reading and mathematics. The most effective interventions include peer‐assisted instruction and small‐group instruction by adults. These have substantial potential to decrease the achievement gap.
What is this review about?
Low levels of mathematics and reading skills are associated with a range of negative outcomes in life, including reduced employment and earnings, and poor health. This review examines the impact of a broad range of school‐based interventions that specifically target students with or at risk of academic difficulties in Grades K‐6. The students in this review either have academic difficulties or are at risk of such difficulties because of their background.
Examples of interventions that are included in this review are peer‐assisted instruction, using financial and non‐financial incentives, instruction by adults to small or medium‐sized groups of students, monitoring progress, using computer‐assisted instruction, and providing coaching to teachers.
Some interventions target specific domains in reading and mathematics such as reading comprehension, fluency, number sense, and operations, while others also focus on building different skills, for example, meta‐cognition and social‐emotional learning.
The review looks at whether these interventions are effective in improving students’ performance on standardised tests of reading and/or mathematics.
What studies are included?
In total, 607 studies are included in this review. However, only 205 of these were of sufficiently high methodological quality to be included in the analysis. Of these, 175 are from the United States, 10 from Sweden, 7 from the United Kingdom, 3 from the Netherlands, 2 from Australia, 2 from Germany, 2 from New Zealand, and 1 each from Canada, Denmark, Ireland, and Israel.
Do targeted school‐based interventions improve reading and mathematics outcomes?
Yes. High‐quality evidence shows that, on average, school‐based interventions aimed at students who are experiencing, or at risk of, academic difficulties, do improve reading and mathematics outcomes in the short term.
What type of intervention is the most effective?
Two instructional methods stand out as being particularly and consistently effective. Both peer‐assisted instruction and small‐group instruction by adults showed the largest (short‐term) improvements in reading and mathematics. Other instructional methods showed smaller improvements however, there is substantial variation in the magnitude of these effects.
Are positive effects sustained in the longer term?
Follow‐up outcomes measured more than three months after the end of the intervention pertain almost exclusively to studies examining small‐group instruction and reading. There is evidence of fadeout but positive effects are still reported up to 2 years after the end of intervention. Only five studies measured intervention effects after more than 2 years.
What do the findings of the review mean?
School‐based interventions in Grades K‐6 can improve reading and mathematics outcomes for students with or at risk of academic difficulties. In particular, the evidence shows that using peer‐assisted instruction and small‐group instruction are two of the most effective approaches that schools can implement. These interventions make a real difference in the achievement gap for at risk students.
At the same time, we need more research to better understand why interventions work better in some contexts compared with others. We also need to know more about the long‐term effects of interventions, and of interventions implemented in other countries than the United States. Furthermore, there are fewer studies of mathematics interventions than reading interventions.
How up‐to‐date is this review?
The review authors searched for studies up to July 2018.
BACKGROUND
Description of the condition
International research has consistently shown that low academic achievement during primary school increases the risk of school dropout, and additionally decreases prospects of secondary or higher education (Berktold et al., 1998; Ensminger & Slausarcick, 1992; Finn et al., 2005; Gardnier et al., 1997; Goldschmidt & Wang, 1999; Randolph et al., 2004; Winding et al., 2013). Entering adulthood with a low level of education is associated with reduced employment prospects as well as limited possibilities for financial progression in adult life (De Ridder et al., 2012; Johnson et al., 2010; OECD, 2012; Scott & Bernhardt, 2000). Furthermore, adults with higher levels of educational attainment are more likely to live longer, show higher levels of civic engagement, and exhibit greater satisfaction with life (OECD, 2010a, 2012). Conversely, low levels of education are negatively correlated with numerous health‐related issues and risk behaviours such as drug use and crime, which have serious implications for the individual as well as for society (Berridge et al., 2001; Brook et al., 2008; Feinstein et al., 2006; Horwood et al., 2010; Sabates et al., 2013).
Overall, in the member countries of the Organisation for Economic Co‐operation and Development (OECD), almost one in five of all youth between 25‐34 years of age have not earned the equivalent of a high‐school degree/upper secondary education (OECD, 2013). Moreover, on average across the OECD countries, around 15% of 18‐24 year‐olds are neither employed, nor in education or training (OECD, 2018). The Programme for International Student Achievement (PISA) tests show that on average about 20%–25% of 15‐year‐olds in the OECD countries are not proficient readers (OECD, 2010b, 2016, 2019).1 Likewise, in mathematics, around 20%–25% of students could only manage the lowest level in the PISA test (OECD, 2010b, 2016, 2019).2 These results indicate that a large proportion of students do not obtain sufficient academic skills in school and stands outside the labour market, once they have left school.
Skill differences between groups of students with low and high risk of ending up with academic difficulties appear early and are often present already before primary school. For example, struggling readers tend to be persistently behind their peers from the early Grades (e.g., Elbro & Petersen, 2004; Francis et al., 1996) and early math and language abilities strongly predict later academic achievement (e.g., Duncan et al., 2007; Golinkoff et al., 2019). Low‐income preschool children have more behaviour problems (e.g., Huaqing & Kaiser, 2003) and there is a strong continuity between emotional and behavioural problems in preschool and psychopathology in later childhood (Link Egger & Angold, 2006). Emotional and behavioural problems are in turn linked to lower academic achievement in school (e.g., Durlak et al., 2011; Taylor et al., 2017). Lastly, the gap between majority and minority children on cognitive skills tests is large already when children are 3–4 years old (e.g., Burchinal et al., 2011; Fryer & Levitt, 2013).
The prenatal and early childhood environment appears to be an important factor that keeps students from realising their academic potential (e.g., Almond et al., 2018). Currie (2009) furthermore documented that children from families with low socioeconomic status (SES) have worse health, including measures of foetal conditions, physical health at birth, incidence of chronic conditions, and mental health problems. Immigrant and minority children are often overrepresented among low SES families and face similar risks (e.g., Bradley & Corwyn, 2002; Deater‐Deckard et al., 1998; Morgan et al., 2012).
Family environments also differ in aspects thought to affect educational achievement: low SES families are less likely to provide a rich language and literacy environment (Bus et al., 1995; Golinkoff et al., 2019; Hart & Risley, 2003). The parenting practices and access to resources such as early childhood education and intervention, health care, nutrition, and enriching spare‐time activities also differ between high‐ and low‐risk groups (Esping‐Andersson et al., 2012; Morgan et al., 2012). Low SES parents also seem to have lower academic expectations for their children (Bradley & Corwyn, 2002; Slates et al., 2012), and teachers often have lower expectations for low SES and minority students (e.g., Good et al., 2003; Timperley & Phillips, 2003). Low SES children are also more likely to experience a decline in motivation during the course of primary, secondary, and upper secondary school (Archambault et al., 2010).
The neighbourhoods that students grow up in is another potential determinant of achievement (e.g., Björklund & Salvanes, 2011; Campbell et al., 2000; Chetty et al., 2018). It seems likely that many students in high‐risk groups live in neighbourhoods that are less supportive of academic achievement in terms of, for example, peer support and role models. To get by in a disadvantaged neighbourhood may also require a very different set of skills compared with what is needed to thrive in school, something that may increase the risk that pupils have trouble decoding the “correct” behaviour in educational environments (e.g., Heller et al., 2017).
As indicated by the previous discussion, the group of students experiencing academic difficulties is diverse. It includes for instance students with learning disabilities, students who are struggling because they lack family support, because they have emotional or behavioural problems, or because they are learning the first language of the country they are living in. Some groups of students may not currently have academic difficulties but are “at risk” in the sense that they are more in danger of ending up with difficulties in the future, at least in the absence of intervention (McWhirter et al., 2004). Although being at risk points to a future negative situation, it is sometimes used to designate a current situation (McWhirter et al., 2004; Tidwell & Corona Garret, 1994), as current academic difficulties are a risk factor for future difficulties and having difficulties in one area may be a risk factor in other areas (McWhirter et al., 1994).
After this review of risk factors for academic difficulties, it is worth noting that the life circumstances placing children and youth at risk are only partially predictive. That is, risk factors increase the probability of having academic difficulties, but are not deterministic. As academic difficulties therefore cannot be perfectly predicted and may show up relatively late in a child's life, interventions in early childhood may not be enough and effective interventions during school may be needed to reduce the achievement gaps substantially.
As the test score gaps between high‐ and low‐risk groups remain relatively stable from the early grades, schools do not seem to be a major reason for the inequality in academic achievement (e.g., Heckman, 2006; Lipsey et al., 2012; von Hippel et al., 2018). Further evidence is provided by the seasonality in achievement gaps. In the United States, the gap between high and low SES students tends to widen during summer breaks when schools are out of session (e.g., Alexander et al., 2001; Gershenson, 2013; Kim & Quinn, 2013; although von Hippel et al., 2018, show that this pattern is not universal across risk groups, grades and cohorts). However, the stability of the test score gaps over time also implies that current school practice is not, in general, enough to decrease the achievement gaps. As schools are perhaps the societal arena where most children can be affected by attempts to reduce the gaps, finding effective school‐based interventions for students with or at risk of academic difficulties is a question of major importance.
Information about effective interventions for students with or at risk of academic difficulties is also of significant interest in most countries. This interest has been reflected in increased political initiatives such as the European Union (EU) Strategic Framework for Education and Training (The Council of the European Union, 2009), or comprehensive legislation such as the No Child Left Behind Act from 2001 in the United States (U.S. Congress, 2002; U.S. Department of Education, 2004).
The research on interventions aimed at academic achievement is rapidly growing, and the interventions described in the literature are numerous and very diverse in terms of for example intervention focus, target group, and delivery mode. The current review focused on targeted, school‐based interventions provided to students in kindergarten (K) to Grade 6 (ages range from 5–7 to 11–13, depending on country/state), where academic learning and skill building were the intervention aims. The outcome variables were standardised tests of achievement in reading and mathematics.
In line with the diversity of reasons for ending up with a low level of skills and educational attainment, we included interventions targeting students who for a broad range of reasons were having academic difficulties, or were at risk of such difficulties. We prioritised already having difficulties over belonging to an at‐risk group in the sense that if there was information about for example test scores and grade point averages, we did not require information about at‐risk status. Furthermore, we did not include interventions targeting high‐performing students in groups that may otherwise be at risk.
This review shares the aims, most inclusion criteria, and the search and screening process with another review about interventions for students in Grades 7–12 (Dietrichson et al., 2020). Consequently, some of the sections below are very similar and a reader that has already read that review may want to skip some parts of the background and method sections.
Description of the intervention
We included interventions that were targeted to students with or at risk of academic difficulties (i.e., interventions that were selected or indicated) and aimed to improve the students’ academic achievement. Targeted interventions can be delivered in various settings, including in class (e.g., peer‐assisted instruction interventions), in group sessions (e.g., the READ180 programme), or one‐to‐one. We restricted the settings to school‐based interventions, by which we mean interventions implemented in school, during the regular school year, and in which schools were one of the stakeholders. This restriction excluded for example after‐school programmes, summer camps and summer reading programmes, and interventions involving only parents and families (see e.g., Zief et al., 2006 for a review of after‐school programmes; Kim & Quinn, 2013, for a review of summer reading programmes; and Jeynes, 2012, for a review of programmes that involve families or parents).
We included a wide range of interventions that aimed to improve the academic achievement of students by changing the method of instruction—such as tutoring, peer‐assisted instruction, and computer‐assisted instruction interventions—or by changing the content of the instruction—for instance, interventions emphasising mathematical problem‐solving skills, reading comprehension, and meta‐cognitive and social‐emotional skills. Many interventions involved changes to both method and content, and included several major components. That is, we included interventions based on their aim to improve academic achievement and based on interventions targeting students with or at risk of academic achievement, and not based on the type of components used in the intervention.
Therefore, we excluded interventions that may improve academic achievement as a side effect, but did not have academic achievement as an aim. Examples are interventions where behavioural or social‐emotional problems were the primary intervention aim. However, interventions with behavioural and social‐emotional components may very well have academic achievement as one of their primary aims, and use standardised tests of reading and mathematics as one of their primary outcomes. Such interventions were included.
Universal interventions applied to improve the quality of the common learning environment at school in order to raise academic performance of all students (including average and above average students) were excluded. We also excluded whole‐school reform strategy concepts such as Success for All, as well as reduced class size interventions and general professional development interventions for principals and teachers that did not target at‐risk students. However, we included some interventions with a professional development component, for example, in the form of coaching of teachers during the implementation of the intervention, as long as the intervention specifically targeted students with or at risk of academic difficulties.
How the intervention might work
All the included interventions strove to improve academic achievement for students with or at risk of academic difficulties. However, they did so with different approaches and with diverse strategies of how to create that improvement. This diversity reflects the varying reasons for why students are struggling or are at risk. In turn, the theoretical background for the interventions varied accordingly. It is therefore not possible to specify one particular theory of change or one theoretical framework for this review. Instead, we briefly review three theoretical perspectives that we believe are characteristic for the majority of the included interventions. We then discuss and exemplify how existing targeted interventions may address some of the reasons for academic difficulties mentioned in Sectionin the light of the theoretical perspectives. 2.1
Theoretical perspectives
The reasons why students may be struggling laid out in the previous section are multifaceted, and the theoretical perspectives underlying the included interventions are broad. Nevertheless, three superordinate components are characteristic for the majority of the included programmes:
We emphasise that the following presentation of these three theoretical perspectives is not exhaustive, and, although components are presented as demarcated, they contain some conceptual overlap.
Social learning theory has its origins in social and personality psychology, and was initially developed by psychologist Julian Rotter and further developed especially by Albert Bandura, (1977, 1986). From the perspective of social learning theory, behaviour and skills are primarily learned by observing and imitating the actions of others, and behaviour is in turn regulated by the recognition of those actions by others (reinforcement), or discouraged by lack of recognition or sanctions (punishment). According to social learning theory, creating the right social context for the student can therefore stimulate more productive behaviour through social modelling and reinforcement of certain behaviours that can lead to higher academic achievement.
Cognitive developmental theory is not one particular theory, but rather a myriad of theories about human development that focus on how cognitive functions such as language skills, comprehension, memory and problem‐solving skills enable students to think, act and learn in their social environment. Some theories emphasise a concept of intelligence where children gradually come to acquire, construct, and use cognitive functions as the child naturally matures with age (e.g., Piaget, 2001; Perry, 1999). Other theories hold a more socio‐cultural view of cognitive development and use a more culturally distinct and individualised concept of intelligence that to a greater extent includes social interaction and individual experience as the basis for cognitive development. Examples include the theories of Robert Sternberg (2009) and Howard Gardner (1999).
Pedagogical theory draws on the different disciplines in psychology and social theory such as cognitivism, social‐interactional theory and socio‐cultural theory of learning and development. There is not one uniform pedagogical model, but examples of contemporary models in mainstream pedagogy are concepts such as Scaffolding (Bruner, 2006) and the Zone of Proximal Development (Vygotsky, 1978), which originated in developmental and educational psychology. These notions hold that learning and development emerge through practical activity and interaction. Acquisition of new knowledge is therefore considered to be dependent on social experience and previous learning, as well as the availability and type of instruction. Accordingly, school interventions require educators to interact and organise the learning environment for the student in certain ways to fit the individual student's needs and potentials for development.
Interventions in practice
School interventions affect academic achievement by changing the methods by which instruction is given (instructional methods) and by targeting certain content (the content domain), and many combine several intervention components as well as theoretical perspectives. Examples of instructional methods covered in earlier reviews are tutoring, coaching of personnel, cooperative learning/peer‐assisted instruction, computer‐assisted instruction, feedback and progress monitoring, and incentive programmes (e.g., Dietrichson et al., 2017). Reading interventions directed to younger students often target content domains as phonemic awareness, phonics, fluency, vocabulary, and comprehension (e.g., Slavin et al., 2009). Slavin and Lake (2008) describe differences in elementary school math curricula in terms of how they emphasise domains such as problem solving, manipulatives, concept development, algorithms, computation and word problems. Gersten et al. (2009) used the following domains to divide mathematics interventions into categories: operations (e.g., addition, subtraction, and multiplication), word problems, fractions, algebra, and general math proficiency (or multiple components). Many school interventions have additional goals concerning other aspects of the student's life, such as reducing problematic behaviour of the students (Cheung & Slavin, 2012; Slavin & Lake, 2008; Wasik, 1997; Wasik & Slavin, 1993).
As indicated, many interventions combine theoretical perspectives. For example, interventions such as tutoring and peer‐assisted instruction interventions often have in common that they comprise an eclectic theoretical model that combines components from all three perspectives on learning presented in the previous section. They are comprehensive interventions that rely on mechanisms such as increased feedback and tailor‐made instruction (pedagogical theory), regulation of behaviour by for example rewards or interaction with role models (social learning theory), and development of cognitive functions such as learning how to learn (cognitive developmental theory).
Another way of viewing these and other types of interventions is that they address the differential family and neighbourhood resources of students with high and low risk of academic difficulties. Low‐risk students are more likely to have access to “tutors” all year round, as parents, siblings, and other family members help out with homework and schoolwork. Interventions to change mindsets, increase expectations, and mitigate stereotype threat may also substitute for low‐risk families and teachers already having such expectations or teaching low‐risk students such a mindset. Different types of extrinsic rewards may be a way to bolster motivation, which may be especially important for students whose families place less weight on educational achievement.
Furthermore, if the differences between students with high and low risk of academic difficulties can be understood as a consequence of differential access to a combination of resources, then remedial efforts may need to address several problems at once to be effective. Programmes that combine certain components may therefore be more effective than others. Another reason why it is interesting to examine combinations of components relates to an often suggested explanation for missing impacts: lack of motivation among participants (e.g., Edmonds et al., 2009; Fuchs et al., 1999). It is therefore possible that programmes will be more effective if they, for example, include some form of rewards for participating students, along with other components providing for instance specific pedagogical support.
Why it is important to do the review
In this section, we first discuss earlier related reviews, and then the contributions of our review in relation to the earlier literature. We focus on reviews that, like our review, compared types of interventions in terms of either instructional methods or content domains.
Prior reviews
Prior reviews have in particular covered reading interventions. Slavin et al. (2009) reviewed reading programmes for elementary Grades. They focused on all kinds of programmes and not only programmes for at‐risk or low‐performing students specifically. Wanzek et al. (2006) reviewed reading programmes directed to students in Grades K‐12 with learning disabilities, and Flynn et al. (2012), Inns et al. (2019), Scammaca et al. (2015), Slavin et al. (2011), and Wanzek et al. (2018) reviewed programmes for struggling readers in Grades 5‐9, K‐5, 4‐12, K‐5, and K‐3, respectively.3 These reviews thus covered low‐achieving students, but neither at‐risk students nor areas other than reading. Suggate (2016) reviewed the long‐run effects of phonemic awareness, phonics, fluency, and reading comprehension interventions from preschool up to Grade 7, but did not discern between interventions targeting students with/at risk of and without/not at risk of academic difficulties.
Mathematics interventions were reviewed in Slavin and Lake (2008) and Pellegrini et al. (2018) for general student populations in elementary school. Gersten et al. (2009) examined four types of components of mathematics instruction for students with learning disabilities, but did not include studies for students at risk of math difficulties (or other reasons for difficulties than learning disabilities). Dietrichson et al. (2017) included interventions targeting both reading and mathematics and based inclusion on the share of students with low SES, but did not consider whether students had academic difficulties or not. Fryer (2017) included both math and reading interventions for all types of student groups.4
All reviews that reported an overall effect size found that it was positive. Most also found substantial variation between interventions. Regarding intervention types, we provide a more detailed comparison to our results in Section 6.5 (including reviews focused on a specific intervention type), but to preview that discussion we describe some overarching results here. Among instructional methods, many reviews indicated that one‐to‐one or small‐group tutoring have relatively large effect sizes across both mathematics and reading interventions compared with other intervention types (Dietrichson et al., 2017; Fryer, 2017; Inns et al., 2019; Pellegrini et al., 2018; Slavin & Lake, 2008; Slavin et al., 2009, 2011). Peer‐assisted instruction or cooperative learning interventions also showed relatively large effect sizes in some reviews (Dietrichson et al., 2017; Inns et al., 2019; Pellegrini et al., 2018; Slavin & Lake, 2008; Slavin et al., 2009, 2011), but not in all (Gersten et al., 2009). Computer‐assisted or technology‐supported instruction have typically positive but smaller effect sizes than small‐group and peer‐assisted instruction (Dietrichson et al., 2017; Inns et al., 2019; Pellegrini et al., 2018; Slavin & Lake, 2008; Slavin et al., 2009, 2011).
Gersten et al. (2009) examined some components of mathematics instruction that do not map neatly into the categories used in the current review and some of the others. They found for example most support for explicit instruction, use of heuristics, and curriculum design. Regarding specific math domains, interventions targeting word problems had higher effect sizes than other math domains but not significantly so.
Reviews focusing on short‐term effects across reading domains reported positive effects in general but few reliable differences over reading domains (Flynn et al., 2012; Scammaca et al., 2015; Wanzek et al., 2006). An exception is that reading comprehension interventions were associated with significantly higher effect sizes than fluency interventions in Scammaca et al. (2015), but this difference disappeared when they only considered standardised tests. Suggate (2016) found that comprehension and phonemic awareness interventions showed relatively lasting effects that transferred to non‐targeted skills, whereas phonics and fluency interventions did not (mean follow‐up was around 11 months).
The contribution of this review
Academic difficulties and lack of educational attainment are significant societal problems. Moreover, as shown by the Salamanca declaration from 1994 (UNESCO, 1994), there has for decades been a great interest among policy makers to improve the inclusion of students with academic difficulties in mainstream schooling, and a desire to increase the number of empirically supported interventions for these student groups.
The main objective of this review is to provide policy makers and educational decision‐makers at all levels—from governments to teachers—with evidence of the effectiveness of interventions aimed to improve the academic achievement of students with or at risk of academic difficulties. To this end, we chose a broad scope in terms of the target group and the types of interventions we included. We included studies that measured the effects of interventions by standardised tests in reading and mathematics. The reason is that many interventions are not directed specifically to either subject and outcomes are therefore measured in both (Dietrichson et al., 2017). Including both students with and at risk of academic difficulties in the target group should also decrease the risk of biasing the results due to omission of studies where information about either academic difficulties or at‐risk status is available, but not both. Furthermore, making comparisons over intervention components within one review, rather than across reviews, should increase the possibilities of a fair comparison. For instance, controlling that effect sizes are calculated in the same way, that the definitions of intervention components are consistent, and that moderators are coded in the same way, is easier within the scope of one review than across reviews.
Earlier reviews with a comparable focus on students with or at risk of academic difficulties have included a more narrowly defined target group. Furthermore, their analyses either did not include intervention components together with other moderators in a meta‐regression, or only included very broad categories of instructional methods and content domains. Such analyses risk confounding the effects of intervention components with for example participant characteristics, and precludes testing whether components have significantly different effect sizes. Furthermore, some reviews have coded interventions regarding the instructional methods used, or regarding the type of content taught, and used such indicators in meta‐regressions (e.g., Dietrichson et al., 2017; Gersten et al., 2009; Scammaca et al., 2015). With the exception of Gersten et al. (2009), who included an indicator for word problems alongside instructional methods‐indicators, the analyses did not include both methods, and content domain indicators. They therefore risk confounding instructional methods with content domains.
Lastly, we are not aware of another review that have provided meta‐analytic estimates of medium‐ and long‐term effects specifically for students with or at risk of academic difficulties.
OBJECTIVES
The primary objective of this review was to assess the effectiveness of targeted interventions aimed at improving the academic achievement for students with or at risk of academic difficulties in Grades K to 6.
The secondary objective was to examine the comparative effectiveness of different types of interventions, focusing on instructional methods and content domains. We conducted subgroup and moderator analyses in which we attempted to identify those methods and domains that have the strongest and most reliable associations with academic outcomes, as measured by standardised test scores in reading and mathematics.
The tertiary objective was to explore the evidence for differential effects across participant and study characteristics. We prioritised characteristics that were relevant for all types of interventions.
METHODS
Criteria for considering studies for this review
Types of studies
According to our protocol, included studies should use an intervention‐control group design or a comparison group design (Dietrichson et al., 2016). Included study designs were randomised controlled trials (RCT), including cluster‐RCTs; quasi‐randomised controlled trials (QRCTs), that is, where participants are allocated by means such as alternate allocation, person's birth date, the date of the week or month, case number, or alphabetical order; and quasi‐experimental studies (QES). To be included, QES had to credibly demonstrate that outcome differences between intervention and control groups is the effect of the intervention and not the result of systematic baseline differences between groups. That is, selection bias should not be driving the results. This assessment is included as a part of the risk of bias tool, which we elaborate on in the “Risk of bias” section, and no QES was excluded on this criterion in the screening process. A fair amount of studies within educational research use single group pre–post comparisons (e.g., Edmonds et al., 2009; Wanzek et al., 2006); such studies were however excluded in the screening process due to the higher risk of bias.
Control groups received treatment‐as‐usual (TAU) or a placebo treatment. We found no studies in which the control group explicitly received nothing (i.e., a no‐treatment control), as all students experienced regular schooling. That is, control groups got whatever instruction the intervention group would have gotten, had there not been an intervention. The TAU condition can for this reason differ substantially between studies (although many studies did not describe the control condition in much detail). Eligible types of control groups included also waiting list control groups, which only differed in the time frame in which researchers estimate the effects. That is, students in both waiting list and regular control groups were offered regular schooling but after the students in the waiting list control group had received the intervention, they could no longer be used as controls.
Comparison designs compared alternative interventions against each other. That is, they made it clear that all students get something other than TAU because of the intervention. Effect sizes from such studies are not fully comparable to effect sizes from intervention‐control designs. We therefore planned to analyse comparison designs separately from intervention‐control designs, and use them where they may shed light on an issue, which could not be fully analysed using the sample of intervention‐control studies. However, the number of studies that were, in this sense, relevant was small and we used them only in one analysis of the effects of group sizes in small‐group instruction interventions.
Due to language restrictions in the review team, we included studies written in English, German, Danish, Norwegian, and Swedish. To ensure a certain degree of comparability between school settings and to align TAU conditions in included studies, we only included studies published in or after 1980.
Types of participants
The population samples eligible for the review included students attending regular schools in Grades K‐6, who were having academic difficulties, or were at risk of such difficulties. Students attending regular private, public, and boarding schools were included, and students receiving special education services within these school settings were also included.
We included only studies carried out in OECD countries. This selection made it more likely that school settings and TAU conditions were comparable across included studies. Grades K‐6 corresponds roughly to primary school, defined as the first step in a three‐tier educational system consisting of primary education, secondary education and tertiary or higher education. We included studies with a student population in higher Grades than K‐6 as long as the majority of the students were in Grades K‐6. The age range included differed between countries, and sometimes between states within countries (ages range from 4–7 to 11–13, depending on country/state). Much fewer studies reported the participants’ ages than Grades, which was also our main reason to formulate the inclusion criteria in terms of Grade rather than age.
The eligible student population included both students identified in the studies by their observed academic achievement (e.g., low academic test results, low grade point average or students with specific academic difficulties such as learning disabilities), and students that were identified primarily on the basis of their educational, psychological, or social background (e.g., students from families with low socioeconomic status, students placed in care, students from minority ethnic/cultural backgrounds, and second language learners). We excluded interventions that only targeted students with physical learning disabilities (e.g., blind students), students with dyslexia/dyscalculia, and interventions that were specifically directed towards students with a certain neuropsychiatric disorder (e.g., autism, ADHD), as some interventions targeting such students are different from interventions targeting the general struggling or at‐risk student population (e.g., they include medical treatments like in Ackerman et al., 1991).
Because there was substantial overlap between students that were already struggling and groups considered at‐risk of difficulties in studies found in a previous review (Dietrichson et al., 2017), we chose to include both students with difficulties and students that were deemed at‐risk, or were considered educationally disadvantaged. If the two criteria were inconsistent, we gave priority to students having academic difficulties. For example, we excluded interventions that targeted high‐achieving students from low‐income backgrounds.
Some interventions included other students, who neither had academic difficulties nor were at risk of such difficulties. For example, in some peer‐assisted learning interventions high‐performing students were paired with struggling students. Studies of such interventions were included if the total sample (intervention and control group) included at least 50% students that were either having academic difficulties or were at risk of developing such difficulties, or if there were separate effect sizes reported for these groups.
Types of interventions
We included interventions that sought to improve academic achievement or specific academic skills. This does not mean that the intervention had to consist of academic activities, but there had to be an expectation in the study that the intervention, regardless of the nature of the intervention content, would result in improved academic achievement or a higher skill level in a specific academic task. We however choose to exclude interventions that only sought to improve performance on a single test instead of improving a skill that would improve test scores. For similar reasons, we excluded studies of interventions where students are provided with accommodations when taking tests; for instance, when some students are allowed to use calculators and others not.
An explicit academic aim of the intervention did not per se exclude interventions that also included non‐academic objectives and outcomes. However, we excluded interventions having academic learning as a possible secondary objective. If the objectives were not explicitly stated, we used the presence of a standardised test in mathematics or reading as an indication that the authors expected the intervention to improve academic achievement. We excluded cases where such tests were included but the authors explicitly stated that they did not expect the intervention to improve reading or math skills.
Furthermore, we only included school‐based interventions. That is, interventions conducted in schools during the regular school year with schools as one of the stakeholders. This latter restriction excluded summer reading programmes, after‐school programmes, parent tutoring programmes, and other programmes delivered in the home of students.
Universal interventions that aimed to improve the quality of the common learning environment at the school level in order to raise academic achievement of all students (including average and above average students), were excluded. Interventions such as the one described in Fryer (2014) where a bundle of best practices were implemented at the school level in low‐achieving schools, where most students are struggling or at risk, was also excluded. This criterion also excluded whole‐school reform strategy concepts such as Success for All, curriculum‐based programmes like Elements of Mathematics (EMP), as well as reduced class size interventions.
This criterion also meant that we excluded interventions where teachers or principals receive professional development training in order to improve general teaching or management skills. Interventions targeting students with or at risk of academic difficulties may on the other hand include a professional development component, for example, when a reading programme includes providing teachers with reading coaches. Such interventions were therefore included.
Our protocol contained no criterion for the duration of interventions and we included interventions of all durations. We coded the duration of the interventions and this variable was included as a moderator in some of the analyses.
Types of outcome measures
We included outcomes that cover two areas of fundamental academic skills:
Studies were only included if they considered one or more of the primary outcomes. Standardised tests included norm‐referenced tests (e.g., Gates‐MacGinitie Reading Tests and Star Math), state‐wide tests (e.g., Iowa Test of Basic Skills), and national tests (e.g., National Assessment of Educational Progress, NAEP). If it was not clear from the description of the outcome measures in the studies, we used online sources to determine whether a test was standardised or not. For example, if a commercial test has been normed, this was typically mentioned on the publisher's homepage. However, for older tests it was not always possible to find information about the test from electronic sources. In these cases, we included the test if there was a reference to a publication describing the test, which made it clear that the test had not been developed for the intervention or the study.
We restricted our attention to standardised tests in part to increase the comparability between effect sizes. Earlier related reviews of academic interventions have pointed out that effect sizes tend to be significantly lower for standardised tests compared with researcher‐developed tests (e.g., Flynn et al., 2012; Gersten et al., 2009; Scammaca et al., 2015). Scammaca et al. (2015) furthermore reported that whereas mean effect sizes differed significantly between the periods 1980–2004 and 2005–2011 for other types of tests, mean effect sizes were not significantly different for standardised tests. As researcher‐developed tests are usually less comprehensive and more likely to measure aspects of content inherent to intervention but not control group instruction (Slavin & Madden, 2011), standardised tests should provide a more reliable measure of lasting differences between intervention and control groups.
We excluded tests that provided composite results for several academic subjects other than mathematics and reading, but included tests of specific domains (e.g., vocabulary, fractions) as well as more general tests, which tested several domains of reading or mathematics. Tests of subdomains had significantly larger effect sizes compared with more general tests in Dietrichson et al. (2017). This result may indicate that it may be easier to improve scores on tests of subdomains than on tests of more general skills, or that tests of subdomains may be more likely to be inherent to intervention group instruction. At the same time, it seems reasonable that interventions that target subdomains of reading and mathematics are tested on whether they affect these subdomains. Therefore, we did not want to exclude either type of test, but coded the type of test and used it as a moderator in the analysis. However, to mitigate problems with test content being inherent to intervention and not control group instruction, we did not consider tests where researchers themselves picked a subset of questions from a norm‐referenced test as being standardised. The subset should either have been predefined (as in e.g., the passage comprehension subset of Woodcock‐Johnson Tests of Achievement) or the picked by someone other than the researchers (e.g., released items from the NAEP).
We included all postintervention tests and coded the timing of each test (see “Multiple time points” section).
Search methods for identification of studies
This section describes the search strategy for identifying potentially relevant studies. We used the EPPI reviewer software to track the search and screening processes. A flowchart describing the search process and specific numbers of references screened on different levels can be found in Section. The search documentation, reporting and details relating to the search can be found in the Supporting Information Appendix. 5.1.2 A
Limitations and restrictions of the search strategy
All searches were restricted to publications after 1980. This year was chosen to balance the competing demands of comparability between intervention settings and comprehensiveness of the review. We used no further limiters in the searches.
Electronic database searches
Relevant published studies were identified through electronic searches of bibliographic databases, government and policy databanks. We searched the following electronic resources/databases:
All databases were originally searched from 1st of January 1980 to March 2016. As mentioned, we only included studies published in or after 1980 to ensure a certain degree of comparability between school settings and to align TAU conditions in included studies. We updated the searches in June/July 2018 using identical search strings. Some database searches were not updated in 2018 due to access limitations.
In Supporting Information Appendix, we report the search strings as well as details for each electronic database and resource searched. A
Note that the searches contained terms relating to secondary school, since the search contributed to a review about this older age group (Grades 7–12, see Dietrichson et al., 2020). There is overlap in the literature among the age groups, and in order to rationalise and accelerate the screening process, we decided upon performing one extensive search.
Searching other web‐based resources
We also searched the following national/international repositories and review/trial archives/registries:
Our protocol stated that we should search two trial registries: The Institute for Education Sciences’ (IES) Registry of Randomized Controlled Trials (http://ies.ed.gov/ncee/wwc/references/registries/index.aspx↗), and American Economic Association's RCT Registry (https://www.socialscienceregistry.org↗). We were however unable to search the IES registry as it was not available (last tried 23 July 2018). We have asked IES about availability, but have to date not received a reply. We updated the search of American Economic Association's RCT Registry on 23 July 2018.
Hand search
The following selected journals had the highest frequency of potentially relevant studies based on the initial pilot‐searches during the development of the search string and the protocol:
The search was performed on editions from 2015 to July 2018 (i.e., including an updated search) of the journals mentioned, in order to capture relevant studies recently published and therefore not found in the systematic search.
Grey literature searches
We performed a wide range of searches on the below institutional and governmental resources, academic clearinghouses and repositories for relevant academic theses, reports and conference/working papers. Most of the resources searched for grey literature include multiple types of references. The resources are listed under the category of literature most prevalent in the resource, even though multiple types of unpublished/published literature might be identified in the resource.
Search for Dissertations
Search for Working Papers/Conference Proceedings
Search for Reports
Contacts to international experts
We contacted international experts to identify unpublished and ongoing studies. We primarily contacted corresponding authors of the related reviews mentioned in Section 2.4.1,5 and authors with many and/or recent included studies. The following authors replied: Douglas Fuchs, Lynn Fuchs, Russell Gersten, Nancy Scammaca, Robert Slavin, and Sharon Vaughn. Furthermore, during work with another review about the use of randomised controlled trials in Scandinavian compulsory school, authors were contacted about studies with sometimes overlapping inclusion criteria with the current review (see Pontoppidan et al., 2018).
Citation‐tracking/snowballing strategies
In order to identify both published studies and grey literature we used citation‐tracking/snowballing strategies. Our primary strategy was to citation‐track related systematic reviews and meta‐analyses. 1446 references from 23 existing reviews were screened in order to find further relevant grey and published studies (see Section 2.4.1 and the list in Supporting Information Appendix A, subsection Grey Literature Searches). The review team also checked reference lists of included primary studies for new leads.
Data collection and analysis
Selection of studies
Under the supervision of the review authors, at least two review team assistants independently screened titles and abstracts to exclude studies that were clearly irrelevant. Any disagreement of eligibility was resolved by the review authors. We retrieved studies considered eligible in full text. Two review team assistants then independently screened the full texts under the supervision of the review authors. Any disagreement of eligibility was resolved by the review authors. The review authors piloted the study inclusion criteria with all review team assistants.
Data extraction and management
Two members of the review team independently coded and extracted data from included studies. A coding sheet was piloted on several studies and revised. Any disagreements were resolved by discussion, and it was possible to reach consensus in all cases. We extracted data on the characteristics of participants, characteristics of the intervention and control/comparison conditions, research design, sample size, outcomes, and results. We contacted study authors if a study did not include sufficient information to calculate an effect size. Extracted data was stored electronically, and we used EPPI Reviewer 4, Microsoft Excel, and R as the primary software tools.
Assessment of risk of bias in included studies
We assessed the risk of bias of effect estimates using a model developed by Prof. Barnaby Reeves in association with the Cochrane Non‐Randomised Studies Methods Group. This model is an extension of the Cochrane Collaboration's risk of bias tool and covers risk of bias in non‐randomised studies that have a well‐defined control group. The extended model is organised and follows the same steps as the risk of bias model according to the 2008‐version of the Cochrane Handbook, chapter 8 (Higgins & Green, 2008). The extension to the model is explained in the three following points:
The refined assessment is pertinent when thinking of data synthesis as it operationalises the identification of studies (especially in relation to non‐randomised studies) with a very high risk of bias. The refinement increases transparency in assessment judgements and provides justification for not including a study with a very high risk of bias in the meta‐analysis.
Risk of bias judgement items
The risk of bias model used in this review is based on nine items (see Supporting Information Appendix B: Risk of bias tool for a fuller description). The nine items refer to: sequence generation, allocation concealment, blinding, incomplete outcome data, selective outcome reporting, other potential threats to validity, a priori protocol, a priori analysis plan, and confounders (for non‐randomised studies). As all but the latter follow standard procedures described in the Cochrane Handbook (Higgins & Green, 2011), we focus on the confounding item below.
Confounding
An important part of the risk of bias assessment of effect estimates in non‐randomised studies is how studies deal with confounding factors. Selection bias is understood as systematic baseline differences between groups and can therefore compromise comparability between groups. Baseline differences can be observable (e.g., age and gender) and unobservable to the researcher (e.g., motivation). Included studies use for example matching and statistical controls to mitigate selection bias, or demonstrate evidence of preintervention equivalence on key risk variables and participant characteristics. In each study, we assessed whether the observable confounding factors of age and Grade level, performance at baseline, gender, and socioeconomic background had been considered, and how each study dealt with unobservables.
There is no single non‐randomised study design that always deals adequately with the selection problem. Different designs represent different approaches to dealing with selection problems under different assumptions and require different types of data. For example, differences in preintervention test score levels do not have to be a major problem in a difference‐in‐differences design, where the main identifying assumption is that the trends of the outcome variable in the intervention and control group would not have differed, had the intervention not occurred (e.g., Abadie, 2005). Similar differences in levels would, in general, be more problematic in a matching design as they indicate that the matching technique has not been able to balance the sample even on observable variables. For this reason, we did not specify thresholds in terms of preintervention differences (in say, effect sizes) for when a study has too high risk of bias on confounding.
Importance of prespecified confounding factors
We describe the motivation for focusing on age and Grade level, performance at baseline, gender, and socioeconomic background below.
Development of cognitive functions relating to school performance and learning are age dependent. Furthermore, systematic differences in performance level often refer to systematic differences in preconditions for further development and learning of both cognitive and social character (Piaget, 2001; Vygotsky, 1978). Therefore, to be sure that an effect estimate was a result from a comparison of groups with no systematic baseline differences it was important to control for the students' Grade level (or age).
Performance at baseline is generally a very strong predictor of posttest scores (e.g., Hedges & Hedberg, 2007), and controlling for this confounder was therefore highly important.
With respect to gender it is well‐known that there exist gender differences in school performance (e.g., Holmlund & Sund, 2005). In terms of our primary outcome measures, girls tend to outperform boys with respect to reading and boys tend outperform girls with respect to mathematics (Stoet & Geary, 2013), although parts of the literature finds that these gender differences vanish over time (Hyde et al., 1990; Hyde & Linn, 1988). As there is no consensus around the disappearance of gender differences, we found it important to include this potential confounder.
Students from more advantaged socioeconomic backgrounds on average begin school better prepared to learn (e.g., Fryer & Levitt, 2013). As outlined in Section 2, students with socio‐economically disadvantaged backgrounds have lower test scores on international tests (OECD, 2010c, 2013). Therefore, the accuracy of the estimated effects of an intervention may depend on how well socioeconomic background is controlled for. Socioeconomic background factors were for example parents’ educational level, family income, and parental occupation.
Bias assessment in practice
At least two review authors independently assessed the risk of bias for each included study. Disagreements were resolved by discussion, and it was possible to reach a consensus in all cases. We reported the risk of bias assessment in risk of bias tables for each included study (see Supporting Information Appendicesand). F G
In accordance with Cochrane and Campbell methods we did not aggregate the 5‐point scale across items. Effect sizes given a rating of 5 on any item should be interpreted as being more likely to mislead than inform and were not be included in the meta‐analysis (the items with a 3‐point scale did not warrant exclusion). If an effect size received a rating of 5 on any item (from both reviewers), we did not continue the assessment because, as per our protocol, these effect sizes would not be included in any analysis. We discuss the risk of bias assessment, including the most common reasons for excluding an effect size, in Section. For studies with a lower than 5‐point rating, we used the ratings of the major items in sensitivity analyses. 5.2
A note is warranted for how we assessed some items in practice. Allocation concealment was assessed as a type of second‐order bias in RCTs. If there was doubt or uncertainty about how the random sequence was generated, this automatically carried over to the allocation concealment rating, which was also rated “Unclear”. Similarly, if the sequence generation rating was “High”, as for example in a QES, then the allocation concealment rating was also “High”. RCTs rated “Low” on sequence generation could get a “High” rating on allocation concealment if the sequence was not concealed from those involved in the enrolment and assignment of participants. However, if the randomisation was not done sequentially, this should not present a problem, and allocation concealment in non‐sequentially randomised RCTs were rated “Low”, given that the rating on sequence generation was also “Low”.
Blinding is in practice always a problem in the interventions we included. No included study was double‐blind for example, a standard that is very difficult to attain in an educational field trial. Furthermore, blinding was not extensively discussed in many studies, likely because it is difficult to attain in education interventions (Sullivan, 2011). For these reasons, we did not exclude any effect size due to insufficient blinding and rather than rating all studies that did not explicitly discuss blinding as “Unclear”, we sought to assess how likely it was that a particular group of participants was blind to treatment status. We used the following categories of participants: students in intervention and control groups, teachers, parents, and testers. We assessed the blinding item by the following standard: if all participant groups were likely to be aware of treatment status, we gave the study a rating of 4. If at least one group was likely blind to treatment status, it got a 3, and then we lowered the rating when more groups were blinded.
There were moreover very few studies that reported having an a priori protocol or analysis plan. We did not count hypotheses stated in the study as an a priori analysis plan. The plan should have been published before the analysis took place and we had to be able to find the plan.
This lack of prespecified outcome measures made it difficult to assess selective outcome reporting bias. However, a few studies lacked information regarding all outcomes described in, for example, the methods section of the study. To separate these effect sizes from the ones that did not contain information about a protocol or an analysis plan, we rated the latter ones with 1 (i.e., there was no evidence of selective outcome reporting). This rating should therefore not necessarily be considered as representing a low risk of bias.
Measures of treatment effect
The analysis of effect sizes involved comparing an intervention to control or comparison conditions. We conducted separate analyses for short‐ and follow‐up outcomes. The below sections apply to both types of outcomes, unless otherwise mentioned.
Effect sizes using continuous data
For continuous data, we calculated standardised mean differences (SMDs) whenever sufficient information was available in the included studies. We used Hedges’ g to estimate SMDs, calculated as (Lipsey & Wilson, 2001, pp. 47–49):
whereis the total sample size,the postintervention mean in each group, andthe pooled standard deviation defined as N = + n 1 n 2 X ® s p
Here,anddenotes the raw standard deviation of the intervention and control group. We used covariate‐adjusted means, and the unadjusted posttest standard deviation whenever available. However, most studies did not report covariate‐adjusted means in a way that we could use. We then used the raw means instead (we test whether the studies reporting only raw means have different effect sizes in the sensitivity analysis). We decided to use the postintervention standard deviation, as more studies included this information than the preintervention standard deviation. In the few cases where the postintervention standard deviation was missing, we used the preintervention standard deviation. s 1 s 2
All studies included in the data synthesis, except one, provided information so that we could calculate student‐level effect sizes. For the exception, we used information about intra‐cluster correlations (ICC) from Hedges and Hedberg (2007, table 6, p. 72, pre‐test covariate model for math in Grade 6, which is 0.098) to transform the teacher/class‐level effect size to a student‐level effect size.
Some studies reported an effect size where the mean difference was standardised using the control group's standard deviation (i.e., a Glass's δ) or reported effect sizes calculated with unclear methods (and no other information that we could use). Furthermore, a few studies used the school‐, district‐, or nation‐wide standard deviation to calculate a standardised mean effect size, but did not include information about the respective standard deviation for intervention and control group. We included these effect sizes, and tested the sensitivity to their inclusion in Section 5.4.
Our protocol stated that we would use intention‐to‐treat (ITT) estimates of the mean difference whenever possible. However, very few studies reported explicit ITTs, and the overwhelming majority only reported results for the students that actually received the intervention, rather than all for which the intervention was intended (often because they lacked outcome data for students that left the study). We therefore believe that the estimates are closer to treatment‐on‐the‐treated (TOT) effects and used TOT estimates when both ITTs and TOTs were available.
A few effect sizes are based on tests were low scores denote beneficial effects. We reverse coded these so that positive effect sizes imply beneficial effects of the intervention.
| Coefficient difference | ‐statisticF | df | Valuep | |
|---|---|---|---|---|
| Hypothesis | (1) | (2) | (3) | (4) |
| Peer‐assisted = CAI | 0.31 | 10.65 | 51.76 | 0.002 |
| Peer‐assisted = Coaching personnel | 0.42 | 8.84 | 43.26 | 0.005 |
| Peer‐assisted = Incentives | 0.31 | 8.49 | 33.26 | 0.006 |
| Peer‐assisted = Medium‐group | 0.21 | 4.4 | 34.3 | 0.043 |
| Peer‐assisted = Progress monitoring | 0.49 | 9.28 | 43.48 | 0.004 |
| Peer‐assisted = Small group | 0.06 | 0.91 | 45.58 | 0.345 |
| Small‐group = CAI | 0.26 | 13.51 | 35.75 | 0.001 |
| Small‐group = Coaching personnel | 0.37 | 9.08 | 48.32 | 0.004 |
| Small‐group = Incentives | 0.25 | 8.46 | 27.6 | 0.007 |
| Small‐group = Medium‐group | 0.15 | 3.58 | 25.21 | 0.07 |
| Small‐group = Progress monitoring | 0.43 | 10.45 | 38.39 | 0.003 |
| Fractions = Algebra/Pre‐algebra | 0.53 | 14.15 | 10.18 | 0.004 |
| Fractions = Geometry | 0.33 | 6.12 | 10.21 | 0.032 |
| Fractions = Number sense | 0.39 | 8.12 | 7.16 | 0.024 |
| Fractions = Operations | 0.47 | 10.09 | 6.29 | 0.018 |
| Fractions = Problem solving | 0.29 | 6.21 | 8.95 | 0.035 |
Effect sizes using discrete data
Only two studies exclusively reported discrete outcome measures. We transformed the outcomes into SMDs using the methods described in Sánchez‐Meca et al. (2003) and included them in the analyses together with studies reporting continuous outcomes.
Unit of analysis issues
Errors in statistical analysis can occur when the unit of allocation differs from the unit of analysis. In cluster‐randomised trials, participants are randomised to intervention and control groups in clusters, as when participants are randomised by school. QES may also include clustered assignment of treatment. Effect sizes and standard errors from such studies may be biased if the unit‐of‐analysis is the individual and an appropriate cluster adjustment is not used (Higgins & Green, 2011).
Our protocol stated that we should adjust studies individually using the methods suggested by Hedges (2007). However, of the 61 studies with clustered assignment of treatment, less than a third contained any information about realised cluster sizes, and estimates of the ICC or the within‐cluster and between‐cluster variances (the ICC is the ratio between the between‐cluster and the total variance). Only a handful contained all the necessary information (both realised cluster sizes, and/or the ICC and the variances). We therefore adjusted all studies in a similar way and used an ICC of 0.09, which is very close to the mean of both reading and mathematics taken over Grades K‐6 in the pretest covariate models of tables 6 and 7 in Hedges and Hedberg (2007, pp. 72–73). In the sensitivity analysis, we report both results using unadjusted effect sizes and using a substantially higher ICC (0.3) than in the primary analysis.
| Excl study characteristics | Incl study characteristics | |
|---|---|---|
| Moderator | (1) | (2) |
| One‐to‐one | 0.171 | 0.077 |
| [−0.035, 0.377] | [−0.124, 0.278] | |
| One‐to‐two or three | 0.199 | 0.081 |
| [−0.015, 0.414] | [−0.150, 0.311] | |
| One‐to‐four or five | 0.169 | 0.015 |
| [−0.066, 0.404] | [−0.238, 0.268] | |
| QES | −0.059 | |
| [−0.327, 0.208] | ||
| Math | 0.078 | |
| [−0.095, 0.252] | ||
| General | 0.03 | |
| [−0.155, 0.217] | ||
| Mean Grade | −0.048 | |
| [−0.084, −0.011] | ||
| Duration | −0.003 | |
| [−0.008, 0.003] | ||
| Constant | 0.18 | 0.26 |
| [−0.010, 0.369] | [0.06, 0.456] | |
| Effect sizes | 583 | 583 |
| Clusters | 95 | 95 |
| N | 60,664 | 60,664 |
| Q | 304 | 272.1 |
| I2 | 70.1 | 68.4 |
| τ2 | 0.08 | 0.08 |
Multiple intervention groups and multiple interventions per individual
Studies with multiple intervention groups with different individuals and studies using multiple tests for the same intervention groups were included in the review. To avoid problems with dependence between effect sizes, we used the robust‐variance estimation (RVE) methods developed by Hedges et al. (2010). We used the results in Tanner‐Smith and Tipton (2014) and Tipton (2015) to evaluate if there were enough studies for this method to consistently estimate the standard errors. That is, we report when the adjusted degrees of freedom are below (or close to) 4 in an analysis or for a moderator. See Section 4.3.9 for more details about the data synthesis. We treated multiple interventions over time as one combined intervention and coded the components.
Multiple studies using the same sample of data
We reviewed studies of interventions given to (partly) the same groups of students, but included one estimate of the effect from each sample of data in the meta‐analysis to avoid overlapping samples. We chose the estimate from the intervention that had lowest risk of bias, or contained the most information. See Supporting Information Appendix C: Studies with overlapping samples or lacking information for a summary description of included studies that we did not include in the meta‐analyses for this reason.
Multiple time points
As per our protocol we divided the analysis into:
In addition to the prespecified analyses, we also conducted analyses of effects measured 3.5 months to 1 year, 1–2 years, and more than 2 years after the end of intervention.
Some studies did not contain exact information about measurement timing. We interpreted these effect sizes as short‐term effects unless there was information in the study that indicated that the measurement was conducted more than 3 months after the end of an intervention. Similarly, for follow‐up measures, we coded the measurement as being within 1 year after the end of intervention when it was clear that the measurement was conducted more than three months after end of intervention but the specific timing was not reported.
If studies tested the same students two or more times within a period, then we used only the measurement closest in time to the end of intervention. That is, if a study tested students at 12 and 18 months after the end of the intervention, we only used the 12 months‐test in the analysis of effect sizes measured between 1 and 2 years after end of intervention.
As we found relatively few long‐term effects and the study variation was limited in ways we describe further below, the examination of heterogeneity and moderator analysis focused on the short‐term effects.
Dealing with missing data
We assessed missing data and attrition rates in the individual studies using the risk of bias tool. Studies had to permit a calculation of a numeric effect size for the outcomes to be eligible for inclusion in the meta‐analysis. Where studies had missing summary data, such as missing standard deviations, we derived effect sizes where possible from, for example, F ratios, t values, χ2 values and correlation coefficients using the methods suggested by Lipsey and Wilson (2001). If these statistics were also missing, we asked the study investigators if they could provide us with the information. We were unable to retrieve information from 24 studies. These studies were included in the review but excluded from the meta‐analysis (see Supporting Information Appendix C: Studies with overlapping samples or lacking information for a summary description).
Many studies did not provide data about all moderators. See Section 4.3.10 and Table 1 for information about missing moderator data. As the number of included studies limited the number of moderators we could include in the analysis, we focused on moderators that had no missing data, were relevant for all types of studies, and were not highly correlated with other moderators.
In a sensitivity analysis, we used multiple imputation to assign values to moderators with relatively low levels of missing information. We defined “relatively low” as missing in less than 20% of interventions. We confined the sensitivity analyses further to moderators that were relevant to all intervention types (e.g., the number of sessions and hours per week is not relevant for incentive and progress monitoring interventions). We constructed imputed data sets using the 3.6.0 version of the mice package in R (first developed by Van Buuren & Groothuis‐Oudshoorn, 2011), which uses chained equations to predict missing values. We averaged our results over five imputed data sets using the methods described in Rubin (1996) and used the predictive mean matching method to assign missing values.
| Study characteristics | k | I | n | Meani | SDi | Rangei |
|---|---|---|---|---|---|---|
| Study context | ||||||
| % performed in the United States | 195 | 327 | 1334 | 0.86 | 0.35 | 0–1 |
| Participants | 195 | 327 | 1334 | 280.34 | 1276.94 | 15–21,317 |
| Districts | 147 | 240 | 959 | 2.91 | 5.03 | 1–27 |
| Schools | 177 | 295 | 1213 | 12.18 | 13.92 | 1–147 |
| Study design and implementation | ||||||
| % QES | 195 | 327 | 1334 | 0.07 | 0.26 | 0–1 |
| % Implementation problems | 177 | 285 | 1172 | 0.11 | 0.32 | 0–1 |
| Outcome assessment | ||||||
| % General test* | 195 | 327 | 1334 | 0.09 | 0.28 | 0–1 |
| % Follow‐up test* | 195 | 327 | 1334 | 0.22 | 0.41 | 0–1 |
| Participant characteristics | ||||||
| % Girls | 167 | 290 | 1254 | 44.58 | 10.03 | 0–65 |
| Grade | 195 | 327 | 1334 | 2.36 | 1.83 | 0–7 |
| % Minority | 154 | 271 | 1139 | 65.49 | 27.28 | 0–100 |
| % Low income | 120 | 192 | 802 | 69.48 | 21.11 | 12.1–100 |
| General intervention characteristics | ||||||
| % Mathematics tests* | 195 | 327 | 1334 | 0.18 | 0.38 | 0–1 |
| Duration in weeks | 194 | 326 | 1329 | 23.11 | 18.01 | 1–160 |
| Number of sessions | 177 | 298 | 1280 | 69.08 | 62.14 | 3–400 |
| Hours per week | 178 | 299 | 1286 | 1.95 | 1.49 | 0.22–10 |
| Implemented by school staff | 192 | 320 | 1325 | 0.51 | 0.5 | 0–1 |
| Instructional methods | ||||||
| Coaching personnel | 195 | 327 | 1334 | 0.11 | 0.31 | 0–1 |
| Computer‐assisted instruction | 195 | 327 | 1334 | 0.15 | 0.36 | 0–1 |
| Incentives | 195 | 327 | 1334 | 0.09 | 0.29 | 0–1 |
| Medium‐group instruction | 195 | 327 | 1334 | 0.06 | 0.25 | 0–1 |
| Other method | 195 | 327 | 1334 | 0.02 | 0.15 | 0–1 |
| Peer‐assisted instruction | 195 | 327 | 1334 | 0.15 | 0.36 | 0–1 |
| Progress monitoring | 195 | 327 | 1334 | 0.13 | 0.33 | 0–1 |
| Small‐group instruction | 195 | 327 | 1334 | 0.65 | 0.48 | 0–1 |
| Content domain | ||||||
| Comprehension | 195 | 327 | 1334 | 0.41 | 0.49 | 0–1 |
| Decoding | 195 | 327 | 1334 | 0.51 | 0.5 | 0–1 |
| Fluency | 195 | 327 | 1334 | 0.3 | 0.46 | 0–1 |
| Multiple reading | 195 | 327 | 1334 | 0.55 | 0.5 | 0–1 |
| Spelling and writing | 195 | 327 | 1334 | 0.23 | 0.42 | 0–1 |
| Vocabulary | 195 | 327 | 1334 | 0.26 | 0.44 | 0–1 |
| Algebra and pre‐algebra | 195 | 327 | 1334 | 0.08 | 0.27 | 0‐1 |
| Fractions | 195 | 327 | 1334 | 0.03 | 0.17 | 0–1 |
| Geometry | 195 | 327 | 1334 | 0.07 | 0.25 | 0–1 |
| Multiple math | 195 | 327 | 1334 | 0.24 | 0.43 | 0–1 |
| Number sense | 195 | 327 | 1334 | 0.15 | 0.35 | 0–1 |
| Operations | 195 | 327 | 1334 | 0.15 | 0.36 | 0–1 |
| Problem solving | 195 | 327 | 1334 | 0.1 | 0.3 | 0–1 |
| General academic skills | 195 | 327 | 1334 | 0.04 | 0.19 | 0–1 |
| Meta‐cognitive strategies | 195 | 327 | 1334 | 0.13 | 0.33 | 0–1 |
| Social‐emotional skills | 195 | 327 | 1334 | 0.05 | 0.22 | 0–1 |
| Single component interventions | ||||||
| Coaching personnel | 195 | 327 | 1334 | 0.01 | 0.1 | 0–1 |
| Computer‐assisted instruction | 195 | 327 | 1334 | 0.06 | 0.23 | 0–1 |
| Incentives | 195 | 327 | 1334 | 0.01 | 0.11 | 0–1 |
| Medium‐group instruction | 195 | 327 | 1334 | 0.04 | 0.19 | 0‐1 |
| Peer‐assisted instruction | 195 | 327 | 1334 | 0.07 | 0.26 | 0–1 |
| Progress monitoring | 195 | 327 | 1334 | 0.01 | 0.1 | 0–1 |
| Small‐group instruction | 195 | 327 | 1334 | 0.47 | 0.5 | 0–1 |
| Comprehension | 195 | 327 | 1334 | 0.02 | 0.13 | 0–1 |
| Decoding | 195 | 327 | 1334 | 0.1 | 0.31 | 0–1 |
| Fluency | 195 | 327 | 1334 | 0 | 0 | 0 |
| Spelling and writing | 195 | 327 | 1334 | 0 | 0 | 0 |
| Vocabulary | 195 | 327 | 1334 | 0 | 0.06 | 0–1 |
| Algebra/pre‐algebra | 195 | 327 | 1334 | 0 | 0 | 0‐1 |
| Fractions | 195 | 327 | 1334 | 0 | 0.06 | 0–1 |
| Geometry | 195 | 327 | 1334 | 0 | 0 | 0 |
| Operations | 195 | 327 | 1334 | 0.01 | 0.11 | 0–1 |
| Number sense | 195 | 327 | 1334 | 0.03 | 0.18 | 0–1 |
| Problem solving | 195 | 327 | 1334 | 0 | 0.06 | 0–1 |
| General academic skills | 195 | 327 | 1334 | 0 | 0.06 | 0–1 |
| Meta‐cognitive strategies | 195 | 327 | 1334 | 0 | 0 | 0 |
| Social‐emotional skills | 195 | 327 | 1334 | 0.01 | 0.08 | 0–1 |
Assessment of heterogeneity
Heterogeneity was assessed with χ2 (Q) test, and the I2, and τ2 statistics (Higgins et al., 2003). In Supporting Information Appendix L, we also provide the prediction intervals for the main effects and subgroup analyses (this analysis was not included in our protocol).
Assessment of reporting biases
Reporting bias refers to both publication bias and selective reporting of outcome data and results. Bias from selective reporting of outcome data and results is one of the main items in the risk of bias tool.
To examine possible publication bias, we used funnel plots, Egger's test (Egger et al., 1997), and tested whether studies published in scientific journals had different effect sizes compared with other studies. We used the R package metafor to conduct these tests, and the restricted maximum likelihood (REML) estimation procedure with the Knapp and Hartung adjustment of standard errors (Viechtbauer, 2010; this procedure was recommended by e.g., Langan et al., 2019).
The simulation results in Pustejovsky and Rodgers (2019), published after our protocol, indicated that the original Egger's test often reject the null hypothesis of no asymmetry at higher rates than the chosen level of statistical significance (i.e., the Type I errors are inflated). We therefore also conducted a version of the “Egger sandwich”‐test suggested by Rodgers and Pustejovsky (2020), which had good Type I properties in their analyses. For the Egger sandwich‐test, we used the same RVE procedure as in the primary analysis (see next section for more details) and simply added a measure of the precision of each effect size, equal to (N/n1n2)0.5, to the respective estimating equation.
Data synthesis
We conducted the overall data synthesis in this review when effect sizes were available. Effect sizes coded with a very high risk of bias (score of 5 on any item in our 5‐point scale) were not included in the data synthesis. The primary analysis had the following steps. We described summary and descriptive statistics of the intervention‐level characteristics, and the risk of bias assessment, and then performed analyses divided by measurement timing (end‐of‐intervention or follow‐up), which corresponded to our first objective for the review. We then explored heterogeneity across instructional methods and content domains (corresponding to the second objective), and other study characteristics (corresponding to the third objective). We describe these subgroup and moderator analyses in the next section.
We used the RVE procedure in the R command robumeta (Fisher et al., 2017) in all our analyses. The RVE procedure allowed us to simultaneously include all effect sizes from each study and avoid problems with dependence between effect sizes for estimation and to calculate robust standard errors (Hedges et al., 2010). We used the random‐effects model weighting scheme option, as it seemed most likely that the effects of the included interventions were not the same across studies, but follow a distribution. A fixed effects model would therefore be less appropriate in our case (e.g., Borenstein et al., 2009). As there were many more effect sizes from studies conducting several tests on the same samples than effect sizes from studies reporting results from independent samples, we chose the correlated effects model instead of the hierarchical effects model in robumeta.
The RVE procedure requires an initial estimate, ρˆ, of the correlation between tests within the same study. We used ρˆ = 0.8 (as e.g., Dietrichson et al., 2017; Hedges et al., 2010; Wilson et al., 2011). We report 95% CIs throughout the analysis (i.e., “statistically significant” denote p < .05) and used the small sample adjusted standard errors and degrees of freedom suggested by Tipton (2015) to calculate CIs. We reported when the adjusted degrees of freedom were close to or below 4, as the results in Tanner‐Smith and Tipton (2014) and Tipton (2015) indicate that the standard errors are not reliable below this level.
Despite that the RVE procedure may have some disadvantages in terms of estimating heterogeneity parameters (see Tanner‐Smith et al., 2016), we chose to use the same framework for all analyses in order to make sure that disparate results were not caused by using different statistical models. We provide a sensitivity analysis for our main results in the “Publication bias” section and in Supporting Information Appendix J: Forest plots by intervention component, where we used study‐level effect sizes (a simple average of the effect sizes in each study) and the REML procedure in the R package metafor in the analysis (Viechtbauer, 2010).
Subgroup analysis and investigation of heterogeneity
One of the main objectives of the review was to assess the comparative effectiveness of intervention components. We therefore performed subgroup and moderator analysis to attempt to identify the characteristics of interventions and study methods that were associated with effect sizes. We again used the RVE procedure in robumeta and reported 95% CIs for regression coefficients. We performed two types of analyses: single‐factor subgroup analysis, in which we estimated RVE models including only an intercept on samples of effect sizes defined by the subgroup of interest, and multiple meta‐regression analyses, in which we estimated RVE models including additional moderators besides the intercept. Although both types of analyses are regression‐based, we sometimes refer to the latter as just “meta‐regressions” to ease the reading. Below we describe the variables we used to define subgroups and as moderators, and thereafter a roadmap for the analysis, which includes a discussion of the advantages and disadvantages with the different types of analyses.
Most included moderators were coded as indicator variables (most variables are natural indicators, e.g., whether the study design was an RCT or not). Continuous variables were mean‐centred to facilitate interpretation. Our protocol specified the following types of moderators:
It is important to note that the number of included studies and the number of effect sizes were not large enough to include all coded moderators in one meta‐regression (see Supporting Information Appendix D: Coding scheme for a description of all coded variables). In line with the objectives of the review and our protocol, we therefore focused the analysis of subgroups and heterogeneity on instructional methods and content domains. These components are substantive features of interventions that for example teachers and principals can affect, in contrast to other moderators (e.g., participant characteristics may be more difficult to affect for a school). They were also more often (quasi‐)experimentally manipulated in studies than other moderators in our sample. They may therefore be less likely to be confounded with other, omitted moderators. However, we want to emphasise that the moderator and subgroup analysis estimate the associations between moderators and effect sizes, and may not capture the causal effects (Thompson & Higgins, 2002).
To further reduce the number of moderators, we first excluded moderators with very low variation (i.e., for which nearly all observations have the same value) or where information was missing from studies. We also excluded moderators that were not relevant for all intervention types (e.g., there is no number of sessions in an intervention that provide students with incentives to read a certain number of books).
We characterised the included interventions using two general categories of treatment modalities or intervention components: instructional method and content domain. As described in our protocol, the components were not fully prespecified, but developed and adapted during the coding process. We used previous reviews and author‐reported classifications in included studies as a starting point, and an iterative process to construct component categories. Below, we describe the coded components by treatment modality, and how we used these components to develop the moderators we included in the analyses. Note that interventions often contained more than one component and they were coded in all component categories they contained. The categories below are therefore not mutually exclusive.
We only coded that an intervention included a component when it was clear from the study that the component was used. For example, Torgesen et al. (2007) write that their intervention contained extensive “professional development and support” but do not mention whether teachers were coached or not. Therefore, we did not code the interventions in this study in the coaching of personnel‐category.
Some studies examined the effects of the same programme (e.g., READ180, PALS). In some cases, the same programme was described differently across the studies. As implementations may differ, we used the descriptions provided in the studies as much as possible and the same programme may thus be coded in different categories. In other cases, information about for example group sizes was lacking from a study and we then inferred a likely group size from information about the programme in other studies of the same programme.
Instructional method
The instructional method‐categories describe the method of delivering the intervention; that is, the contrast between how the intervention group and the control group were instructed. Many interventions contained more than one instructional method. In these cases, we have coded the intervention in all categories.
Coaching of personnel
Interventions in this category included programmes that provided teachers or other school personnel with coaches. Coaching of personnel hired by the research team, for example, ongoing training of college students acting as tutors, was not coded in this category. Furthermore, note that this component did not include professional development interventions that seek to develop more general teaching or management skills, as such interventions were never targeted to at‐risk students in our sample. The coaching in this category was mainly connected to the implementation of a specific reading or mathematics programme.
Computer‐assisted instruction (CAI)
This category indicated whether the intervention, or parts of the intervention, involved computers, tablets, or similar devices in the instruction of students. Computer assistance to teachers was not coded in this category. For example, in Fuchs et al. (1997) teachers get feedback from curriculum‐based measurement implemented on a computer. However, it is only the test that is taken on the computer, which is not used in the instruction of students. Consequently, we did not code this intervention in the CAI‐category (but in the progress monitoring‐category).
Incentives
Incentive programmes intended to increase the academic performance of students were included in this category. The incentives were not always monetary, non‐financial incentives were also included. Examples included interventions where the incentive component was the only component, for instance, students were paid to perform on interim assessments or to improve general achievement. Most interventions combined incentives with other components.
Peer‐assisted instruction
We separated between adult‐led instruction and peer‐assisted instruction. We defined peers as students in Grades K‐12. Interventions such as cross‐age tutoring where fourth graders tutored second graders were thus coded as peer‐assisted instruction (for both tutors and tutees if results were reported for both groups). If on the other hand college students acted as tutors to primary school students, the intervention was coded as adult‐led small‐group instruction (see below for description). We coded the exact group size, if available in the studies, but to keep the number of moderators down, we used a single moderator for the peer‐assisted instruction category in the main analysis. Most studies used small‐groups like pairs.
Progress monitoring
This category included interventions that added a specific progress monitoring component, where teachers received detailed information about the students’ development. Note that for example small‐group interventions of all kinds are also likely to contain increased feedback and in a sense increased (informal) progress monitoring. These interventions were not automatically coded in this category. Interventions had to add an extra component of progress monitoring, such as using curriculum‐based measurements (CBM) during the intervention, to be coded here. Few studies used progress monitoring as the only instructional method, it was almost always combined with other methods.
Medium‐ and small‐group instruction
As mentioned, we separated between adult‐led instruction and peer‐assisted instruction, and these two categories included adult‐led instruction. In some interventions, instruction was given in class, and not divided into smaller groups (this was, or was very likely to be, the same type of instruction given to the control group, so we did not create a moderator for this group size). We coded the exact group size whenever available but quite a few studies did not provide exact information about group size (but reported for example a range). We coded interventions without specific information about group size in the most likely category, given other information in the study or based the coding on information from other studies of the same intervention (e.g., there are several studies of READ180).
Because of the missing data, and to keep the number of moderators down, we created two moderators of adult‐led instruction in groups of at most five students (small‐group instruction) and groups of 6–20 students (medium‐group instruction). A smaller group usually meant that the information was included, and more exact, and we therefore coded interventions where it was clear that the instructional group were smaller than a whole class, but not clear whether it was small‐group or medium‐group instruction in the latter category. Some interventions vary the group size during the intervention. That is, they used both small‐group and medium‐group instruction. We then coded them in both categories. We used these two moderators in most of our analyses, but we also examined finer categories and used comparison designs that assigned group size (quasi)‐experimentally in some analyses.
Lastly, we created a category called “other method” that included interventions that were not coded in any of the above categories. There were two types of interventions making up this category: in a few cases, there was no difference in how the intervention and control group was instructed. Either only the content differed, or the intervention group was just provided extra instruction time. As many studies did not provide information about how much instruction time the intervention and the control group got in a certain area, the latter case was difficult to assess systematically for all interventions. Therefore, we did not create a separate category for extra instruction time.
Content domain
The content domain describes the area targeted by the intervention and the material taught. Interventions often follow the curriculum but put more focus on certain domains. The difference between the treatment and control group was therefore less sharp for content domains compared with instructional methods. We divided these components into reading, mathematics, and other areas. For reading, we used the following categories:
Comprehension
Reading comprehension interventions focused on the understanding and learning from text. Reading comprehension is described by the National Reading Panel (2000) as an active process where interaction between the text, the author, and the reader results in the understanding or meaning making of the text. The RAND Reading Group defines comprehension as “the process of simultaneously extracting and constructing meaning through interaction and involvement with written language” (RAND Reading Study Group, 2002, p. 720).
Decoding
Decoding interventions focused on the translation of print to speech. This category included for example word identification, word study, and word reading interventions. We also included interventions in this category that taught phonological awareness, phonemic awareness, and phonics. Such skills are often thought to be precursors to efficient decoding.
Fluency
We defined fluency as the ability to read orally with speed, accuracy, and proper expression (The National Reading Panel, 2000). Interventions in this category aimed for example to improve the ability to read in a “smooth and effortless” manner (Allinder et al., 2001).
Spelling & Writing
Some interventions included spelling and writing training, which, while not strictly a reading skill, we thought were related enough (and was also tested with standardised reading tests).
Vocabulary
This category included interventions focused on increasing the number of words a student knows. We also included the teaching of “sight words”—that is, frequently occurring words with spelling‐to‐sound irregularities (Castles et al., 2018)—in this category.
In addition to these single domains, we also coded a multiple reading domain category. Besides interventions focused on more than two of the above subdomains, this category included interventions that focused on reading in general but did not explicitly mention any subdomains. We interpreted these interventions as implicitly targeting more than one reading domain.
For math interventions, we found interventions targeting the following categories: 6
Algebra/pre‐algebra
Algebra and pre‐algebra interventions focused on, for example, the basics of equations, and graphs.
Fractions
Fraction interventions taught the concept of fractions and how to manipulate them.
Geometry
Geometry refers to the study of, for example, shapes, sizes, and positions.
Number sense
Number sense, or number knowledge, interventions targeted basic skills such as counting, number recognition, number relations (bigger and smaller, before and after), and number set operations (e.g., Dyson et al., 2015).
Operations
This category included for example training in addition, subtraction, multiplication, and more generally, computational skills.
Problem solving
Interventions coded in this category trained students in solving word problems. For example by teaching students the structural features underlying different problem types (e.g., Fien et al., 2016).
As for reading, we coded a multiple mathematics domains category, which included interventions explicitly covering more than one domain as well as more general math interventions.
Finally, we coded three categories to characterise interventions targeting other areas instead of, or together with (subdomains of) reading and mathematics.
Meta‐cognitive strategies
Meta‐cognitive strategies and self‐regulation interventions aimed to help students think about their own learning more explicitly, and develop strategies for such learning, including managing one's own motivation towards and engagement with learning. The intention was often to give pupils a repertoire of strategies to choose from during learning activities, including study skills and learning how to learn. In comparison to the next domain, social‐emotional skills, the skills trained in this category were more focused on the individual student, and less on the relations to other students or school staff.
Social‐emotional skills
Interventions in this category focused on improving academic achievement through e.g., improving social skills, and mitigating problematic behaviour. They thus had a more relational focus compared with meta‐cognitive interventions.
General academic skills
This category included studies without a particular content domain or a more general academic focus than just reading and math. As the authors studying such interventions still included a standardised test in reading or math, we interpreted the authors as expecting the intervention to improve achievement in these subjects.
Other intervention characteristics
When coding other intervention characteristics, we used information about the intervention group, if available. If information was not available on the intervention group level, we used joint intervention and control group information, and then higher levels, such as Grades and schools. We treated information on levels higher than schools, such as school districts, as missing. Some studies included only information about intervention characteristics given in a range. In these cases, we used the midpoint of that range. Below, we first describe the moderators used in the analysis in more detail and then describe moderators that we coded, but for different reasons did not include in the analysis.
Study context
We coded the country where the information was performed. When information was missing, we made an assessment based on e.g., where the authors were based at the time of the study, and on the mentioning of country‐specific reforms like No Child Left Behind in the United States. We reported the number of participants, schools and districts involved in the study.
Study design characteristics
We coded a moderator indicating whether the study was a QES. We found only one QRCT and coded this study in the QES category. That is, the reference category is RCTs. We coded whether implementation was monitored by the researchers in some way, whether problems were mentioned, and if so, what type of problems that was mentioned. Some problems mentioned by more than one study were low attendance, that implementers had low quality of implementation or low motivation, and that some in the control group might have received (some of) the intervention. In a sensitivity analysis, we used an indicator equal to one if implementation problems were explicitly mentioned.
Effect size measurement
We calculated effect sizes on the basis of different types of tests, which may cause heterogeneity. We therefore coded the content domains of the tests. In the meta‐regressions, we included one moderator indicating whether a test was general, in the sense that it covered two or more subdomains. We furthermore coded four moderators that relate to the calculation of effect sizes, which we used for sensitivity analyses. The first indicated whether we had to use the raw means to calculate the effect size. Glass's δ indicated whether the SMD was standardised with the control group's standard deviation. This was the case in some studies that did not include information about the pooled standard deviation, and rather than excluding them, we tested whether our results are sensitive to their inclusion. Another moderator indicated effect sizes where it was unclear exactly how the effect size was calculated (we believed that it was either Hedges’ g, Cohen's d, or Glass's δ). We also used an indicator equal to one if the SMD had been standardised with a standard deviation from a super‐population (e.g., Grade, district, or state) instead of the intervention and control group, or if the number of included schools, districts or regions was larger than the intervention median.
Both standardisation with a super‐population and including more schools, districts, and regions may imply that the variance in the sample could be larger and effect sizes mechanically smaller, as the study included a possibly more varied group than other studies (see e.g., Lipsey et al., 2012). As there were few studies that standardised with a super‐population, we chose to make one variable for these two related problems.
Participant and sample characteristics
We measured the gender distribution by the share of girls. We coded both age and Grade (minimum, maximum, and mean for both) but the information about age, as well as minimum and maximum for both variables, were missing for far more interventions. We therefore focused on the mean Grade, and used the information about the mean age and the school system in the few studies missing Grade information to estimate a mean Grade. If we only had information about a range, we used the midpoint of that range. That is, if a study reported students in kindergarten to Grade 4, we used Grade 2 as the mean. Outcomes were normally measured in the same Grade that the intervention was performed, but in some cases interventions spanned one or more Grades. The Grade variable we used refers to the Grade in which (most of) the intervention was implemented. To standardise the start of primary school across countries, we used the United States as a starting point and coded kindergarten as Grade 0. That is, if the average grade in a study is 0.5, then half the treatment group were kindergartners and half first‐graders. We recoded Year or Grade 1 to Grade 0 for studies conducted in countries like Australia, New Zealand, and the United Kingdom, where Year/Grade 1 denotes the first year of primary school.
In addition to the mean Grade, we coded the share of minority students (defined as not being part of the largest population group in a country) and the share of students from low‐income families, which was almost always measured as the share of students with free‐ or reduced price meals. As the criteria for getting free‐ or reduced price meals differ between countries, the latter variable was difficult to define consistently across countries.
Dosage
We coded three variables related to the dosage of an intervention. We measured duration in weeks. We used 40 weeks for a school year, and consequently 20 weeks for a semester. We measured the frequency of an intervention by the total number of sessions, and the intensity by the total number of intervention hours per week. For these dosage‐variables we coded both intended and received dosage. However, many studies either lacked information about the intended or received dosage. We used received dosage as a starting point, and added intended in the cases were the received number was missing.
Implementers
We used information about who implemented the intervention to develop an indicator for interventions implemented by school staff (e.g., teachers, special education teachers, coaches, teaching assistants, and teacher aides) with other types of implementers. Peer‐assisted instruction interventions, where peers deliver the instruction, where indicated in this category if school staff facilitated the instruction. The reference category included other type of instructors and implementers such as researchers, college students, or adult volunteers.
Our protocol mentioned and we coded information from several moderators, which we in the end could not use. As we believed they were unlikely to be representative of the literature, we did not report descriptive statistics about these moderators. The main reason for not using them was lack of information: For example, only 3 studies included in the meta‐analysis provided information about parental occupation and 14 about parental education. The share of students speaking the majority language as a second language (e.g., English language learners) was only included in 59 studies.
Two cases where we lack information are, we believe, especially important to highlight: First, the instruction given to the control group differed between interventions. Control group instruction was nearly always some form of TAU, but it was difficult to separate different TAUs from each other as the information was not detailed enough. We were therefore unable to create moderators measuring the quality of the control group instruction. We describe a way to test the sensitivity to differences in control group instruction further below. Second, we coded information about the target group of the intervention, but it was difficult to use this information to develop a moderator measuring the severity of academic difficulties. One reason was that many studies did not provide sufficient detail about how they assessed difficulties or at‐risk status. A few studies specifically targeted students diagnosed with learning disabilities, which is an option for a moderator measuring academic difficulties. However, other studies also included some learning disabled students and many more did not include information about the share of learning disabled students. Due to the small number of studies and the unclear contrast, we refrained from using learning disability as a moderator.
For other moderators, there was very little variation: almost all studies were conducted only in schools, and few had target groups that were not defined in terms of having academic difficulties. That is, few studies defined the target group purely in terms of for instance the students’ SES. We coded whether implementers received training before the intervention, but if this was not the case, it almost always meant that it was a researcher or someone affiliated with the research team who performed the intervention. The information is therefore overlapping with the variables measuring who implemented the intervention. We coded whether the control group was a waitlist design, but it was often not explicitly mentioned whether the control group got the intervention after the intervention group.
Roadmap for the subgroup and moderator analysis
Our investigation of the heterogeneity of the short‐term effects have five main parts:
Sensitivity analysis
We performed the following sensitivity analyses.
Effect size measurement
We tested sensitivity to measurement of effect sizes in the following ways: We included four moderators that indicated whether we used the raw means to calculate the effect size, standardised the SMDs with the control group standard deviation (i.e., Glass's δ), the standardisation was unclear, or the effect size was standardised with a standard deviation from a super‐population.
Outliers
We examined the distributions of effect sizes for the presence of outliers and the sensitivity of our main results by methods suggested by Lipsey and Wilson (2001): trimming the distribution by dropping the outliers and by winsorizing the outliers to the nearest non‐outlier value.
Clustered assignment of treatment
We tested sensitivity to clustered assignment of treatment by the methods described in Section. 4.3.5
Moderators with missing values
We used multiple imputation to account for missing values in some moderators with relatively low rate of missing values, as described in Section. 4.3.6
Control group progression
We coded a description about the control group condition, but it proved difficult to develop an informative moderator based on this coding (due to missing or insufficient information in a large share of studies). To gauge whether differences in the quality of instruction given to the control group may explain some of the effect size heterogeneity, we calculated the control group progression from pre‐ to posttest and divided by the control group posttest standard deviation in studies that included this information. We then (a) tested whether this control group “effect size” was heterogeneous across studies, and (b) used a meta‐regression to examine whether intervention components explained some of the heterogeneity. If there is heterogeneity and some of this heterogeneity is systematic across components, then we risk confounding the intervention components with the quality of control group instruction. However, not all studies included the necessary information. More progression may also mean that it is easier to improve the achievement of the students in that sample (intervention and control group) than in others, which would not necessarily bias our estimates. This test should therefore be interpreted with caution.
Risk of bias
We used the items with numerical ratings from the risk‐of‐bias assessment to examine if methodological quality was associated with effect sizes. The items with non‐numerical ratings are not relevant for all types of studies, and there was also low variation in their ratings. Because the items are categorical variables, we recoded them to indicator variables. For blinding, incomplete outcome reporting, and other bias, we contrasted effect sizes given a rating of 4 or unclear to those given lower ratings than 4. For the selective outcome reporting item, we contrasted those rated 1 (a large majority) with those given higher ratings or an unclear rating. The confounding item is only relevant for QES and captures features that are included in the other bias item for RCTs. It was therefore difficult to compare QES and RCTs in the same analysis. As we included few QES in the meta‐analysis (17) and our ratings of the confounding item exhibited relatively low variation (about 70% of the effect sizes received a rating of 4), we omitted the QES from this sensitivity analysis.
Publication bias
Lastly, we examined publication bias by the methods described in Section. 4.3.8
Deviations from protocol
The search strategy in our protocol listed “Education Research Complete” among the databases. However, at the time of the search, we no longer had institutional access to this database and did not include it either in our original search or in the updated search. Furthermore, when we wrote protocol, we did not have access to the Teacher Reference Center database. During the search process, we did have access, so we chose to search and include the database in the review.
Due to lack of institutional access, we did not search British Education Index, FRANCIS, Dissertation & theses A&I, CBCA Education, and Australian Education Index in the updated search.
According to our protocol, we were supposed to contact international experts to identify studies and give them the list of included studies. We thought it would be advantageous to involve experts earlier in the process and asked them about relevant studies before our screening process was completed. Therefore, they did not receive a list of included studies. As mentioned earlier, we could not search the Institute for Education Sciences’ Registry of Randomized Controlled Trials, as the webpage was shut down at the time of both the original and updated search.
Our protocol stipulated that we would use ITT estimates whenever available. Very few studies reported ITT estimates, however. The estimates reported were in our view closer to TOT estimates and we therefore decided to use the TOT estimates also in studies that reported both ITT and TOT estimates.
Our protocol did not mention the calculation of prediction intervals. As prediction intervals provide a measure of the dispersion of effect sizes (i.e., how effect sizes vary across populations; Borenstein et al., 2017), we reported the prediction intervals for the main effects and subgroup analysis as a complement to the other heterogeneity statistics (see Supporting Information Appendix L).
After the publication of our protocol, Pustejovsky and Rodgers (2019) showed that that the original Egger's test often reject the null hypothesis of no asymmetry in funnel plots at higher rates than the chosen level of statistical significance. To check weather our results were sensitive to this problem, we also conducted a version of the “Egger sandwich”‐test suggested by Rodgers and Pustejovsky (2020).
Lastly, we conducted the analyses in R instead of Stata due to better availability of packages for meta‐analysis.
RESULTS
Description of studies
Results of the search
Figure 1 displays the results of the search process. The total number of potentially relevant records was 24,414 after excluding 187 duplicates (database search: 17,444; grey literature: 3014; citation tracking: 1509; author contacts: 576; hand search: 1024; trial registries and others: 847).
We screened all records based on title and abstract. Of the ones that we did not exclude, 201 records were not retrievable in full text. Older reports and dissertations were overrepresented among the records we could not retrieve. We screened the remaining 4247 retrievable records in full text. A large number of studies were not relevant for this review due to the Grade of the participating students. Studies that were relevant except for the grade of participating students were included in a review covering Grades 7–12 (Dietrichson et al., 2020). In total, 607 studies met the inclusion criteria for this review and we extracted data from these studies.
Flowchart of the search and screening process
Included studies
We did not include all 607 studies in the meta‐analyses. We were unable to retrieve sufficient information from 24 studies/study authors to calculate an effect size. We excluded 17 studies, which used samples that overlapped with other included studies and had either higher risk of bias or contained less information. 104 studies were not included in the meta‐analyses due to their study design. These studies used comparison designs that contrasted two alternative interventions and the contrast was not recurring in enough studies for meta‐analysis to be meaningful (one study included both an intervention‐control contrast and a comparison design and we counted it among the studies included in the meta‐analysis). We did not include 257 studies in the meta‐analysis due to the risk of bias assessment. All eligible outcomes in these studies had, in our view, too high risk of bias (for more details, see further below).
Furthermore, some of the remaining 205 (202 intervention‐control and 3 comparison designs) studies contained overlapping samples, but included for example information about short‐term and follow‐up outcomes. These were all included in some meta‐analyses, but never in the same. We treated studies that reported short‐ and follow‐up outcomes from the same intervention in separate papers as one study in the analysis (i.e., when clustering the standard errors). This definition left us with 195 included clusters of studies in the analysis of intervention‐control studies (none of the comparison design studies had this problem).
We discuss the results of the risk of bias assessment further in Section 5.2. See also the Risk of bias tables in Supporting Information Appendix F for details of the assessment for effect sizes that we included in the meta‐analysis, as well as those that we deemed had too high risk of bias (Supporting Information Appendix G). We describe the comparison designs in the Supporting Information Appendix E: Description of comparison designs. For more information about studies with overlapping samples or that lacked sufficient information for the calculation of an effect size, see the Supporting Information Appendix C: Studies with overlapping samples or lacking information.
The data we coded for the 195 (clusters of) intervention‐control studies and the comparison design studies included in the meta‐analyses are included in Supporting Information Appendicesand. Below we describe the characteristics of the 195 intervention‐control studies, which we based most of the analyses on. H I
Figure 2 displays the included studies by publication year. Most studies were published in the last 10–15 years, 28 included studies were published before the year 2000 and only 7 before 1990. It is therefore unlikely that we missed many relevant studies by restricting our searches to studies published after 1980. Figure 3 shows the 195 studies by the mean Grade of participating students during the implementation of the intervention. There were more studies with participants in kindergarten and, in particular, first Grade than in the higher Grades. The mean Grade was 2.4.
Table 1 contains descriptive statistics of the intervention‐control studies included in the meta‐analysis. Many studies contained more than one intervention, the effects of which may have been tested with more than one standardised test from which we calculated the effect sizes. We denoted the number of studies that provided information about a certain characteristic with k, the number of interventions with i, and the number of effect sizes with n (note that these are not necessarily unique study populations, as some studies with more than one intervention group used only one control group). As most characteristics vary on the intervention level, we averaged – with three exceptions marked with *—over interventions to calculate the mean, standard deviation, and range. For example, in a study with two interventions where each intervention and control group take two tests, we averaged by intervention over the two tests. These averages are the basis for means, standard deviations, and ranges in Table 1 (which is why there is an i subscript in the table). There were in total 195 studies, 327 interventions, and 1334 effect sizes.
Included interventions were to a large extent conducted in the United States (86%). The remaining interventions were from Australia (0.6%), Canada (0.9%), Denmark (0.3%), Germany (1.2%), Ireland (0.3%), Israel (0.3%), Netherlands (2.1%), New Zealand (2.1%), Sweden (3.7%) and the United Kingdom (2.1%). The mean number of participants, schools, and districts were 280, 12, and 3, respectively, but sample sizes varied quite widely and many studies were small. Most included study designs were RCTs, only 7% of interventions were QES. A small share, 11%, reported having some form of implementation problems, and a large majority of the tests (91%) tested a single reading or mathematics domain, that is, they were not general in our terminology. In 22% of the interventions, follow‐up tests were conducted (i.e., students were tested more than 3 months after the end of intervention).
Participants were more likely to be boys (45% were girls), and a majority were minority (65%) and low‐income students (70%). Information about minority and, in particular, low‐income students was relatively often missing. Note that we often based the share of low‐income students on the share receiving free‐ and reduced price lunches in studies from the United States, which is not necessarily directly comparable to low‐income variables used in other countries (e.g., free school meals in the United Kingdom).
The effects of interventions were more often examined using reading tests, 18% used a mathematics test. Note that the separation of effect sizes was made based on the test, not the subject targeted, as there were several interventions that used tests in both reading and mathematics (i = 19), or used a composite reading and mathematics tests (i = 2). This is also a main reason why we do not separate results into reading and mathematics interventions to start with (we will return to this issue in the analysis of heterogeneity).
The mean duration was about 23 weeks, and the mean frequency and intensity equalled 69 sessions and 2 h per week. The range was wide for all three variables measuring intervention dosage and note that we lack information for quite a few interventions regarding frequency and intensity (for 28 and 27 interventions, respectively). Among half of the interventions were implemented by school staff.
Among the instructional methods, small‐group instruction is clearly the most studied method. Around 65% of all interventions included small‐group instruction. The shares otherwise ranged from 2% in the other method‐category to 15% in the CAI and peer‐assisted instruction category.
The proportions of interventions targeting reading domains were more similar, from fluency being targeted in 30% of all interventions to decoding that was targeted by 51% of all interventions. Most interventions (55%) targeted multiple reading areas. Fewer interventions targeted mathematics and the shares ranged from 3% for fractions to 15% for number sense and operations. Targeting more than one domain was common also for math interventions, 24% of interventions targeted more than one math domain. Regarding the non‐math and non‐reading domains, meta‐cognitive strategies was taught by 13%, social‐emotional skills by 5%, and general academic skills by 4%.
The fact that combinations of both instructional methods and content domains was relatively common imply that there are few single component interventions. As shown in the lower part of the table, it is only small‐group instruction that have been studied alone in a large number of studies. Small‐group instruction was the only instructional method in 50% of all interventions. Among the content domains, single domain interventions were rare.
Excluded studies
Due to the large number of studies screened in full text, we were unable to describe all excluded studies. To exemplify how we applied the inclusion criteria, this section describes studies that met all but one of our criteria.
The included study designs contrasted intervention and control groups, or alternative interventions, to estimate effects. Compton et al. (2012) examined 129 first Grade children who were unresponsive to classroom reading instruction and were randomly assigned to 14 weeks of small‐group intervention. They defined nonresponders and responders using a standardised test of word identification, and then used logistic regression models to examine which information predicted the group students ended up in. As the study did not compare an intervention group to a control group, we excluded it.
We included interventions that sought to improve academic achievement or specific academic skills. We however excluded interventions that used accommodations to improve results on tests or only sought to improve test‐taking skills, like Scruggs and Mastropieri (1986). We also excluded interventions that primarily aimed to improve student attendance, like the intervention examined by Guryan et al. (2017). Although improving attendance may also improve student achievement, it need not do so on standardised tests. Furthermore, in groups of low attenders, improving attendance in the intervention group compared with the control group changes the composition of the students who take standardised tests, making it difficult to evaluate whether academic skills have improved (e.g., because the programme may affect student achievement through other channels than just attendance, an instrumental variables analysis need not work, even if the programme is randomly assigned).
Interventions had to be targeting students with or at risk of academic difficulties to be included. Al Otaiba et al. (2011) examined an intervention where teachers received help to differentiate reading instruction in kindergarten based upon students’ ongoing assessments of language and literacy skills. Although the intervention included students’ from at‐risk groups (e.g., students who qualified for special education) and some schools were Title I and Reading First schools (indicating that a relatively large share of the students were from disadvantaged groups), we excluded the study. The programme did not specifically target students with or at risk of academic difficulties, but all students in the participating schools, including high‐performing students.
Interventions should be school‐based to be included, meaning that they were conducted in school during regular school‐hours and semesters. Kim (2006) studied a voluntary summer reading intervention. Although some reading lessons took place in during the spring semester, most of the intervention period was outside the regular school year. We therefore excluded the study.
Studies had to test the effects of interventions using standardised tests in reading and mathematics. Fortner (1986) tested intervention effects with the Test of Written Language (TOWL). TOWL is standardised but primarily a test of writing. However, it includes a subtest of vocabulary but the students’ vocabularies are assessed based on their writing products. Therefore, we thought this subtest was more of a writing than a reading assessment, and excluded the study.
Risk of bias in included studies
We excluded 257 studies that met our inclusion criteria from the meta‐analyses because we rated all relevant effect sizes as having too high risk of bias. That is, our assessment was that all effect sizes in the 257 studies were more likely to mislead than inform the analysis. Studies in which some effect sizes but not all were rated too high risk of bias are counted among the 205 studies included in the meta‐analysis.
It is important to note that we assessed the risk of bias of the effect sizes that met our inclusion criteria and, in turn, the effect estimates we could use to calculate effect sizes, not all effect estimates in a study. That is, there may well be estimates with low risk of bias in the studies that we excluded from the meta‐analysis because they did not fit our inclusion criteria, or because they were reported in a way that we could not use them. For example, to calculate effect sizes from certain model‐based estimates require information about correlations between the variables in the model (Lipsey & Wilson, 2001). Such correlations were rarely included in the studies. Then we could only use the raw means to calculate effect sizes, which, at least in QES, often have too high risk of bias. Furthermore, contrasts between, for example, at‐risk and not‐at‐risk students or at‐risk students and national norms, are often informative but not comparable to effects estimated with a control group of at‐risk students. Thus, a too high risk of bias rating is not a synonym for a low‐quality study.
Included studies for which we rated all effect sizes as having too high risk of bias were disproportionally dissertations (33% compared with 7% among the intervention‐control studies included in the meta‐analysis), older studies (mean publication year is 2004 compared with 2007), and QES (84% of the excluded studies were QES). The most common reasons for giving a too high risk of bias rating were:
We excluded 40 RCTs. RCTs were for example excluded because randomisation was compromised and there was inadequate control for confounding (17 studies), because only one unit was assigned to the intervention or control group (10 studies), because the studies reported results for a subset of included tests or students (6 studies), or because of large‐scale attrition (4 studies). We excluded the 3 remaining studies for more idiosyncratic reasons. We listed the reasons for giving a rating of too high risk of bias by study (for both RCTs and QES) in Supporting Information Appendix G: Studies with a too high risk of bias rating. We reported the main reason per study, but note that there were cases with more than one reason for too high risk of bias ratings.
Risk of bias of effect sizes included in the meta‐analysis
Figure 4 shows the distribution of the assessments for intervention‐control effect sizes included in the meta‐analysis by the items in the risk of bias tool. See Supporting Information Appendix F: Risk of bias in studies included in the meta‐analysis for a description of the ratings by study and item.
Few RCTs reported how they generated the random sequence used to assign students to intervention and control groups, only 7% of effect sizes were given a low‐risk assessment (QES have high risk by default on this item and therefore also on allocation concealment). More generally, the procedure of randomisation was often not described in detail. In almost all cases where the random sequence generation was described and was adequate, the allocation was likely concealed, as the randomisation was not done sequentially.
All studies had problems with the blinding of treatment status. No effect size received a rating of 1 on this item and very few studies provided an explicit discussion about blinding. There was some variation between effect sizes though: For around 66% of effect sizes, there was no indication that any participant group was blind to treatment status. About 32% of effect sizes had one group that was blind to treatment status (usually the persons performing the tests), and in 2%, several groups were likely blinded. The ratings for this item vary to some extent also within studies, as some studies, for example, used both tests performed by persons outside the study (e.g., state‐wide tests) and by involved, non‐blinded, study personnel.
The distribution of assessments for incomplete outcome reporting was more mixed. Only 5% had a high risk of bias rating on this item and almost all studies provided information. We rated a large majority (74%) of effect sizes to be free of selective reporting, but this does not mean that the studies followed a prespecified protocol and analysis plan. The figure omits the items examining if the study followed an a priori protocol and analysis plan, as just two studies mention a protocol and three an analysis plan explicitly written before conducting the analysis. Torgerson et al. (2011) was the only study for which we could retrieve both a protocol and an analysis plan. Lastly, about 17% of effect sizes were rated as having a high risk of bias for the other bias item.
The confounding item was only assessed for the 19 QES; 75% of effect sizes from these studies received a rating of 4. That is, a high risk of bias. Only 7% of effect sizes from QES received a rating of 1 or 2.
In sum, the included effect sizes and studies have relatively often a high risk of bias, but there is also variation. We return to the sensitivity of our results to different part of the risk of bias assessment in Section. 5.4
Summary of risk of bias items for effect sizes included in the meta‐analysis
Effects of interventions
Overall short‐term and medium‐ to long‐term effects
This section presents the results from the robust‐variance estimation of the overall short‐term effects—that is, from the end of intervention to 3 months after—and effects with a longer follow‐up period.
We included 1030 effect sizes, 189 clusters, and 206,186 student observations8 in the analysis of short‐term effects. Eleven individual studies did not provide results from a short‐term test. The weighted average short‐term effect size was positive and statistically significant (ES = 0.30, CI = [0.25, 0.34]). This effect size corresponds to a 58.4% chance that a randomly selected score of a student who received the intervention is greater than the score of a randomly selected student who did not (a null effect would imply a 50% chance, see e.g., Ruscio, 2008, for a conversion formula). The Q‐statistic was 797.3 (p < .01), the τ2 0.067 and the I2 was 76.4. All three heterogeneity measures therefore indicated substantial heterogeneity.9
Figure 5 displays the distribution of short‐term effect sizes. The figure underscores that the effect sizes are heterogeneous. Although most effect sizes are centred around the mean (the red line), there are examples of very large positive effect sizes as well as large negative effect sizes, indicating substantial heterogeneity.
Most studies did not report follow‐up effect sizes: there were 195 effect sizes from 27 studies measured more than 3 months after the end of intervention, which included 19,902 student observations. The weighted average follow‐up effect size was positive and significant (ES = 0.27, CI = [0.17, 0.36]). This effect size corresponds to a 57.5% chance that a randomly selected score of a student who received the intervention is greater than the score of a randomly selected student who did not. The Q‐statistic was 47.8 (p < .01), the τ2 0.03, and the I2 was 45.6, which, although lower than for the short‐term effects, indicated that there was some systematic variation in the effect sizes and significant heterogeneity.10
The average follow‐up effect size was almost as large as the average short‐term effect size. The reason is not only that studies measured outcomes very close to 3 months after the end of intervention. Exploratory analyses revealed that the average effect size measured between 4 and 12 months after the end of intervention was ES = 0.26 (CI = [0.15, 0.37]) and between 12 and 24 months was ES = 0.17 (CI = [−0.03, 0.37]). These analyses included 22 and 9 studies, respectively. That is, the effects were still substantial also for the longer follow‐up periods. There were only 5 studies measuring effects after more than 24 months (ES = 0.11, CI = [−0.14, 0.36]) and the adjusted degrees of freedom fell below 4. This result was therefore unreliable.
It is possible that the follow‐up measurements are mainly confined to interventions that were successful in the short‐term (we found no example of a study with a protocol that detailed a follow‐up measurement in advance). We found an average short‐term effect size of ES = 0.40 (CI = [0.24, 0.55]) among the 22 studies that provided both a short‐term measure and at least one follow‐up measure. That this effect size was larger than the effect size for all studies is an indication that successful interventions are more likely to be examined with follow‐up tests. In turn, this may explain part of the similarity between the short‐term and follow‐up average effect sizes in our sample.
Studies with follow‐up measurements were also different in two other ways. All but five studies used small‐group instruction (i.e., student groups of five or below) and all but eight studies tested effects only on reading measures (two of the eight studies tested both math and reading). Among the exceptions, two studies used peer‐assisted instruction, one study used groups of max eight students, one study used CAI and incentives, and one study changed only the content domain. Thus, evidence of medium‐ to long‐term effects pertains almost exclusively to small‐group instruction interventions that examined the effects on reading measures. Confining the analysis of follow‐up effects (>3 months after end‐of‐intervention) to studies using small‐group instruction yields an ES = 0.28 (CI = [0.16, 0.39]) and confining the analysis further to include only interventions using small‐group instruction and testing effects on reading measures yields an ES = 0.27 (CI = [0.12, 0.42]).
Summarising the analysis thus far, we found evidence of reasonably large and statistically significant short‐term and follow‐up effects. All measures indicated substantial heterogeneity of the short‐term effects, which included a mix of intervention types. The evidence for the follow‐up effects pertains almost exclusively to studies examining small‐group instruction and to effects on reading measures, and we found few studies that have examined effects more than two years after the end of intervention. Because of the small number of studies, and the few instructional methods and content domains used in the follow‐up studies, we focused the subgroup analysis and investigation of heterogeneity in the following section on the short‐term effects.
Distribution of short‐term effect sizes
Results of the subgroup analysis and investigation of heterogeneity
The previous analyses indicated substantial heterogeneity of the short‐term effects. This section examines if we can explain some of this heterogeneity using subgroup and moderator analysis. The analysis follows the five‐step roadmap laid out in Section: 4.3.10
| (1) | (2) | (3) | (4) | (5) | (6) | (7) | (8) | (9) | (10) | (11) | (12) | (13) | (14) | (15) | (16) | (17) | (18) | (19) | (20) | (21) | (22) | (23) | (24) | (25) | (26) | (27) | |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 1 | 1 | ||||||||||||||||||||||||||
| 2 | –0.07 | 1 | |||||||||||||||||||||||||
| 3 | 0.06 | –0.06 | 1 | ||||||||||||||||||||||||
| 4 | –0.09 | –0.06 | –0.04 | 1 | |||||||||||||||||||||||
| 5 | –0.05 | –0.03 | –0.03 | –0.03 | 1 | ||||||||||||||||||||||
| 6 | –0.13 | 0.03 | 0.13 | 0 | –0.04 | 1 | |||||||||||||||||||||
| 7 | 0.11 | 0.07 | 0.03 | –0.06 | –0.04 | 0.05 | 1 | ||||||||||||||||||||
| 8 | –0.35 | –0.01 | –0.17 | –0.31 | –0.19 | –0.51 | –0.13 | 1 | |||||||||||||||||||
| 9 | 0.06 | 0.17 | 0.02 | –0.03 | –0.09 | –0.07 | –0.04 | 0.11 | 1 | ||||||||||||||||||
| 10 | –0.05 | 0 | –0.16 | 0.05 | –0.02 | –0.30 | –0.01 | 0.23 | 0.32 | 1 | |||||||||||||||||
| 11 | –0.17 | 0.13 | 0.08 | 0.02 | –0.10 | −0.08 | –0.07 | 0.2 | 0.53 | 0.29 | 1 | ||||||||||||||||
| 12 | –0.14 | 0.07 | –0.03 | 0 | –0.09 | –0.13 | –0.08 | 0.24 | 0.67 | 0.47 | 0.6 | 1 | |||||||||||||||
| 13 | 0.02 | –0.03 | –0.11 | 0.01 | –0.03 | –0.17 | –0.02 | 0.12 | 0.16 | 0.37 | 0.09 | 0.43 | 1 | ||||||||||||||
| 14 | −0.03 | 0.14 | –0.04 | 0.12 | –0.04 | –0.17 | –0.11 | 0.07 | 0.52 | 0.29 | 0.42 | 0.45 | 0.09 | 1 | |||||||||||||
| 15 | 0 | –0.03 | –0.01 | –0.05 | −0.03 | –0.01 | 0.02 | 0 | –0.23 | –0.30 | –0.18 | –0.31 | –0.13 | –0.15 | 1 | ||||||||||||
| 16 | –0.05 | 0.01 | 0.01 | –0.04 | –0.02 | –0.05 | 0.06 | 0.01 | –0.15 | –0.20 | –0.13 | –0.21 | –0.09 | –0.10 | 0 | 1 | |||||||||||
| 17 | 0.09 | –0.07 | –0.07 | –0.06 | 0.01 | –0.01 | 0.01 | –0.07 | –0.24 | –0.32 | –0.20 | –0.33 | –0.14 | –0.16 | 0.5 | 0.36 | 1 | ||||||||||
| 18 | 0.06 | −0.03 | 0.21 | −0.05 | 0.05 | 0.04 | 0.1 | −0.14 | −0.43 | −0.57 | −0.35 | −0.55 | −0.25 | −0.27 | 0.53 | 0.34 | 0.57 | 1 | |||||||||
| 19 | 0.1 | −0.06 | 0.19 | −0.08 | −0.01 | 0.01 | 0.07 | −0.08 | −0.34 | −0.44 | −0.27 | −0.45 | −0.19 | −0.21 | 0.12 | 0.3 | 0.42 | 0.66 | 1 | ||||||||
| 20 | 0.11 | −0.05 | 0.16 | −0.09 | −0.01 | 0 | 0.11 | −0.10 | −0.37 | −0.48 | −0.30 | −0.49 | −0.21 | −0.23 | 0.33 | 0.34 | 0.47 | 0.76 | 0.72 | 1 | |||||||
| 21 | 0.01 | −0.07 | 0.16 | −0.06 | −0.03 | 0 | −0.02 | 0.01 | −0.25 | −0.33 | −0.20 | −0.34 | −0.15 | −0.15 | 0.72 | −0.01 | 0.41 | 0.57 | 0.29 | 0.43 | 1 | ||||||
| 22 | −0.09 | 0.06 | 0.02 | −0.05 | 0.07 | −0.03 | −0.06 | 0.02 | −0.20 | −0.25 | −0.16 | −0.18 | −0.12 | −0.13 | −0.04 | −0.03 | −0.04 | 0.04 | −0.03 | −0.03 | −0.05 | 1 | |||||
| 23 | −0.08 | 0.1 | 0.32 | −0.03 | 0.05 | 0.14 | −0.03 | −0.08 | −0.07 | −0.18 | −0.10 | −0.04 | −0.14 | 0.02 | −0.03 | 0 | −0.04 | 0.03 | 0.07 | 0.03 | 0.03 | 0.2 | 1 | ||||
| 24 | −0.08 | 0.04 | −0.05 | 0.02 | 0.12 | 0.14 | −0.04 | −0.08 | −0.18 | −0.17 | −0.07 | −0.14 | −0.11 | −0.08 | −0.04 | −0.03 | −0.04 | 0.01 | −0.06 | −0.07 | −0.04 | 0.19 | 0.17 | 1 | |||
| 25 | −0.05 | −0.07 | 0.06 | 0.17 | 0.05 | 0.04 | −0.04 | −0.04 | −0.05 | −0.20 | −0.07 | −0.05 | 0.05 | 0.01 | 0.06 | −0.03 | −0.05 | 0.05 | −0.04 | 0 | 0.05 | −0.04 | −0.07 | −0.04 | 1 | ||
| 26 | 0.01 | 0.16 | 0.01 | −0.05 | 0.15 | 0.16 | 0.13 | −0.21 | −0.16 | −0.32 | −0.20 | −0.22 | −0.11 | −0.12 | 0.1 | −0.03 | 0.02 | 0.19 | 0.22 | 0.09 | 0.03 | 0.31 | 0.15 | 0.18 | 0.13 | 1 | |
| 27 | 0.12 | −0.01 | 0.04 | 0.14 | 0.02 | 0.19 | 0.02 | −0.31 | 0.19 | −0.21 | 0.08 | 0.15 | 0.03 | 0.08 | 0.05 | 0.06 | −0.04 | 0.01 | ‐0.19 | −0.12 | 0.04 | 0.04 | 0.13 | 0.14 | 0.05 | 0.11 | 1 |
Math and reading tests and combined intervention components
We first estimated a model including all posttest effect sizes, an intercept representing effect sizes based on reading tests, and an indicator for math tests. The overall effect size for reading tests was 0.28 and on math tests 0.33, and the difference was not significant (math indicator CI = [−0.04, 0.15]). In the rest of this section, we pooled math and reading tests in the analysis.
Figures 6, 7, 8, 9 show weighted average effect sizes and 95% CIs from RVE estimations by instructional method (Figure 6) and by content domain (Figure 7, 8, 9). We derived each effect size from a meta‐regression including just a constant with the outcome variable being the effect sizes from interventions including the component in question. Note that the effect sizes should be interpreted as the weighted average effect size for interventions that included a certain component, not the effect size of that component in isolation (see below for such estimates). Figure 6 indicates that all instructional methods were associated with positive and statistically significant average effect sizes. Peer‐assisted instruction is associated with the largest effect sizes (ES = 0.44, CI = [0.28, 0.61]) and other methods with the smallest effect sizes (ES = 0.12, CI = [0.01, 0.23]).
The average effect sizes of the reading domains are all reasonably close to each other in Figure 7. The effect sizes range from about 0.20 to 0.30, all of which are statistically significant. Figure 8 shows that the average effect sizes of the math domains varied more than the reading domains: from 0.14 for algebra/pre‐algebra domain to 0.47 for fractions (all effect sizes are statistically significant). The effect sizes for interventions targeting meta‐cognitive, social‐emotional, and general academic skills, shown in Figure 9, were also positive. However, the effect size for the social‐emotional domain was not significant.
Table 3 summarises the average effect sizes per component, the CIs, number of studies, effect sizes, student observations, the adjusted degrees freedom, and provides heterogeneity statistics. The adjusted degrees of freedom are above 4 for all components, but are relatively close for components examined in few studies, such as other methods, algebra/pre‐algebra, fractions, geometry, and general academic skills. Although the heterogeneity was reduced compared with the analysis of overall short‐term effect sizes, the average effect sizes for most components were still substantially heterogeneous. Partial exceptions were CAI (τ2= 0.014, I2 = 31.5, Q = 39.4), progress monitoring (τ2 = 0.021, I2 = 44.6, Q = 32.5), algebra/pre‐algebra (τ2 = 0.024, I2 = 47.1, Q = 22.7), and geometry (τ2 = 0.028, I2= 52.6, Q = 19.0), but the Q‐test indicated that heterogeneity was still statistically significant also for these components.
In Supporting Information Appendix J: Forest plots by intervention component, we show forest plots corresponding to the analyses in Table 3. We estimated the forest plots using study‐level average effect sizes and the REML option in the R package metafor (Viechtbauer, 2010). This study‐level analysis provided a robustness check of the RVE procedure and made the forest plots more legible. The results corroborate the results reported here: the effect sizes are in all instances close to those in reported in Table 3. All estimates that are statistically significant in Table 3 are also significant in Supporting Information Appendix J, except for the other methods‐category. Averaging effect sizes by study reduces the heterogeneity. For example, the Q‐test is not significant for CAI, other method, progress monitoring, algebra/pre‐algebra, geometry, and operations in Supporting Information Appendix J. However, most components display a broad range of minimum and maximum values and the prediction intervals include or are at zero for all except CAI, and geometry.
The advantages of this analysis are that more studies can be included in each subgroup and that the effect sizes are more comparable with those estimated in some earlier reviews. As mentioned, there are also a few important drawbacks. If two or more components are included in an intervention, we cannot separately identify the association of any component. As shown in Table 1, several components were always used in combination with at least one other component. Table 2 contains the correlation coefficients between our moderators and indicates that some components are highly correlated, and that few correlations are zero. This risk was therefore pertinent. In the next section, we show estimates from interventions that contained a single instructional method or a single content domain to mitigate this risk.
Subgroup analyses: Weighted average effect sizes and 95% confidence intervals by instructional method
Subgroup analyses: Weighted average effect sizes and 95% confidence intervals by reading domain
Subgroup analyses: Weighted average effect sizes and 95% confidence intervals by math domain
Subgroup analyses: Weighted average effect sizes and 95% confidence intervals other content domains
| Component | Avg. ES | 95% CI lower bound | 95% CI upper bound | K | n | N | Adj.sdf | τ2 | I2 | Q |
|---|---|---|---|---|---|---|---|---|---|---|
| CAI | 0.151 | 0.086 | 0.217 | 28 | 171 | 29,910 | 19.6 | 0.014 | 31.5 | 39.4 |
| Coaching | 0.2 | 0.084 | 0.316 | 21 | 82 | 24,945 | 16.6 | 0.045 | 75.3 | 81.1 |
| Incentives | 0.328 | 0.184 | 0.472 | 19 | 78 | 30,812 | 16.5 | 0.068 | 67.2 | 54.9 |
| Medium‐group | 0.32 | 0.054 | 0.587 | 18 | 62 | 13,684 | 16.2 | 0.169 | 86.8 | 129.2 |
| Other method | 0.116 | 0.006 | 0.225 | 8 | 14 | 47,511 | 6 | 0.064 | 85.7 | 48.9 |
| Peer‐assisted | 0.444 | 0.276 | 0.613 | 32 | 97 | 8725 | 29.6 | 0.168 | 74.9 | 123.4 |
| Progress mon. | 0.173 | 0.071 | 0.274 | 19 | 102 | 23,164 | 11.7 | 0.021 | 44.6 | 32.5 |
| Small‐group | 0.376 | 0.314 | 0.438 | 118 | 756 | 89,295 | 107.4 | 0.088 | 75.3 | 472.9 |
| Comprehension | 0.238 | 0.179 | 0.297 | 73 | 538 | 69,469 | 58.8 | 0.041 | 62.3 | 191.2 |
| Decoding | 0.29 | 0.228 | 0.352 | 92 | 673 | 74,627 | 79.8 | 0.061 | 66.5 | 271.6 |
| Fluency | 0.258 | 0.186 | 0.329 | 56 | 432 | 56,055 | 47.3 | 0.053 | 65.9 | 161.3 |
| Multiple reading | 0.272 | 0.211 | 0.333 | 103 | 685 | 138,453 | 87.5 | 0.057 | 77.1 | 444.9 |
| Spelling/writing | 0.317 | 0.22 | 0.413 | 42 | 276 | 29,953 | 35.5 | 0.062 | 70.1 | 137.2 |
| Vocabulary | 0.2 | 0.143 | 0.258 | 55 | 325 | 52,181 | 42.5 | 0.029 | 58.2 | 129.3 |
| Algebra | 0.148 | 0.007 | 0.289 | 13 | 46 | 12,495 | 8.8 | 0.024 | 47.1 | 22.7 |
| Fractions | 0.501 | 0.146 | 0.857 | 6 | 22 | 4260 | 4.8 | 0.181 | 86.6 | 37.2 |
| Geometry | 0.169 | 0.092 | 0.246 | 10 | 53 | 11,112 | 6.6 | 0.028 | 52.6 | 19 |
| Multiple math | 0.281 | 0.203 | 0.358 | 48 | 149 | 87,328 | 41.6 | 0.061 | 79.4 | 227.9 |
| Number sense | 0.324 | 0.234 | 0.414 | 30 | 96 | 22,335 | 27 | 0.058 | 72.1 | 103.8 |
| Operations | 0.292 | 0.215 | 0.37 | 31 | 112 | 23,740 | 26.4 | 0.041 | 61.4 | 77.7 |
| Problem solving | 0.328 | 0.176 | 0.48 | 17 | 56 | 7399 | 14.3 | 0.057 | 58.5 | 38.5 |
| Gen. academic | 0.213 | 0.05 | 0.375 | 11 | 37 | 23,713 | 8.7 | 0.053 | 77.3 | 44.1 |
| Meta‐cognitive | 0.242 | 0.153 | 0.331 | 31 | 87 | 24,746 | 25 | 0.044 | 73.9 | 114.7 |
| Social‐emotional | 0.241 | ‐0.135 | 0.618 | 13 | 35 | 9186 | 11.6 | 0.216 | 87.3 | 94.3 |
Single instructional method and content domain interventions
Table 4 shows the results for the single instructional method and single domain estimations. Single method interventions examined interventions using only one of our categories of instructional methods. Single domain interventions targeted only one content domain. As some content domains were not studied in isolation in more than one study in our sample, we omitted them from Table 4. By definition, this analysis does not include the multiple domain‐categories. Furthermore, we omitted the other method category from the table, which is a single method (or no method) category already in Table 3.
Few single components were examined in enough studies for this strategy to give reliable results. Coaching of personnel, incentives, and progress monitoring as well as all content domains except decoding and number sense have adjusted degrees of freedom below 4. For the components that the analysis resulted in adjusted degrees of freedom above 4, the average effect sizes are mostly similar to the ones in Table 3. The exception is number sense, where the effect size in single domain interventions is quite a lot larger (0.51 instead of 0.32). Among the content domains, decoding was the only other statistically significant domain besides number sense. Both peer‐assisted instruction and small‐group instruction continued to be associated with large and statistically significant effect sizes (0.38 and 0.37, respectively). CAI had smaller (0.13) but statistically significant positive effects.
For most components, the number of studies also precluded conclusions about heterogeneity. Confining the comments to components examined in more than 2 studies, CAI and peer‐assisted instruction was examined in reasonably many studies (12 and 16), had relatively low τ2 (0.015 and 0.023, respectively) and I2 (37.0 and 19.8, respectively), and had a Q‐statistic that imply that we cannot reject the null hypothesis of no heterogeneity (p = .10 and p = .23, respectively). Medium‐ and small‐group instruction, decoding, and number sense had significant levels of heterogeneity.
This estimation strategy isolates the association between methods/domains and effect sizes better than the previous, but few components were examined in enough single method/single domain studies for this strategy to give reliable results. So far, neither analysis have adjusted for other study characteristics, and not for instructional methods and content domains at the same time, which is what we do in the next section.
| Component (single) | Avg. ES | 95% CI lower bound | 95% CI upper bound | k | n | N | Adj.sdf | τ2 | I2 | Q |
|---|---|---|---|---|---|---|---|---|---|---|
| CAI | 0.128 | 0.018 | 0.239 | 12 | 67 | 10,995 | 7.9 | 0.015 | 37 | 17.5 |
| Coaching | −0.047 | −0.878 | 0.783 | 3 | 4 | 2957 | 1.7 | 0.05 | 68.9 | 6.4 |
| Incentives | 0.046 | −0.086 | 0.178 | 3 | 7 | 22,154 | 1.3 | 0 | 0 | 2 |
| Medium‐group | 0.091 | −0.092 | 0.274 | 10 | 45 | 9219 | 7.4 | 0.044 | 61 | 23.1 |
| Peer‐assisted | 0.387 | 0.257 | 0.518 | 16 | 54 | 4208 | 12.6 | 0.023 | 19.8 | 18.7 |
| Progress mon. | 0.277 | −3.085 | 3.638 | 2 | 4 | 90 | 1 | 0 | 0 | 1 |
| Small‐group | 0.375 | 0.304 | 0.446 | 85 | 538 | 51,445 | 76.7 | 0.084 | 70.6 | 285.4 |
| Comprehension | 0.205 | −0.409 | 0.819 | 3 | 13 | 1504 | 1.7 | 0.021 | 23.6 | 2.6 |
| Decoding | 0.305 | 0.145 | 0.465 | 25 | 111 | 6653 | 22.9 | 0.148 | 66.5 | 71.6 |
| Number sense | 0.51 | 0.143 | 0.877 | 6 | 13 | 2296 | 4.5 | 0.11 | 82.1 | 27.9 |
| Operations | 0.165 | −0.149 | 0.479 | 3 | 10 | 3653 | 1.7 | 0.049 | 54.5 | 4.4 |
Multiple meta‐regressions on the full sample of short‐term effect sizes
Table 5 shows results from four multiple meta‐regressions in which we included indicators for each instructional method and content domain, and, in some regressions, additional study characteristics. The table displays the coefficient estimates and 95% CIs (directly beside the coefficient estimates). To retain enough (adjusted) degrees of freedom, we included only study characteristics without missing information: the mean Grade (mean centred) and indicators for QES, general tests, and mathematics test.
Column 1 presents results from a specification including indicators for all components but no constant and no additional study characteristics. This specification reports the total marginal association between each component and effect sizes, conditional on all other components. Among the instructional methods, there are two statistically significant associations with effect sizes: peer‐assisted instruction (β = .41) and small‐group instruction (β = .36). Medium‐group instruction, CAI, and other method have insignificant associations slightly above 0.1, while all other methods have smaller positive or negative associations (all insignificant). Fractions is the only content domain with a significant positive association with effect sizes (β = .33). The coefficients for most other domains are close to zero, only spelling and writing is above 0.1. All component indicators have adjusted degrees of freedom above 4 and the only component close to 4 is fractions (adjusted degrees of freedom = 4.3), which, as seen in Table 3, has been examined in only 6 studies.
In column 2, we added a constant and the study characteristics without missing values, and excluded the indicators for other method, multiple reading, and multiple math to get a reference category and avoid multicollinearity. The total marginal association for a specific component is therefore the sum of the constant and the coefficient on the component. However, as the constant is virtually 0, the total marginal association can be read off the individual coefficients.
Adding study characteristics did not change the results much. Spelling and writing has a positive and statistically significant association and algebra/pre‐algebra a negative and statistically significant association, but the size of the total marginal associations are close to those in column 1. Both the size and significance of the other instructional methods and content domains are close to the results in column 1 with the partial exception of number sense, for which the association decreases by around 0.1 (it is still insignificant).
Mean Grade is the only statistically significant study characteristic. The coefficient indicates that interventions in higher Grades have smaller effect sizes (about −0.03 per Grade). The results for the other study characteristics indicate that effect sizes from QES were not significantly different from RCTs, general tests did not have significantly different effect sizes from tests of subdomains, and effect sizes based on math tests were not significantly different from those based on reading tests. All included variables have adjusted degrees of freedom above 4 (with fractions again being closest to 4, with a value of 5.3).
In columns 3 and 4, we report results where we separated the analysis into effect sizes based on reading tests (column 3) and based on mathematics tests (column 4) and leave out all content domains. Although we found no strong indication that effect sizes were different between the two subjects in column 2, associations for specific instructional methods may still differ across subjects. Indeed, there are some differences: both CAI and coaching of personnel have positive associations with effect sizes based on reading tests, and negative associations with effect sizes based mathematics tests. Only the coefficient for coaching on math tests is significant. Note however that the constant is much larger in column 4 compared with column 3, meaning that the differences across the specifications between the total marginal associations for CAI and coaching of personnel are smaller. The association between medium‐group instruction and effect sizes also shifts from positive and relatively large for reading tests to negative and relatively large for mathematics tests. However, neither of the two coefficients is statistically significant and the adjusted degrees of freedom falls below 4 in column 4. Otherwise, all included variables in both columns 3 and 4 have adjusted degrees of freedom above 4. Lastly, the coefficients on peer‐assisted instruction and small‐group instruction are reasonably similar across the specifications, large, and statistically significant.
As there were few content domains with large, stable, and significant associations with effect sizes and few instructional methods with enough studies/effect sizes, we refrained from using a similar specification for content domains. The systematic variation in effect sizes, as measured by the I2, was between 62 and 66 throughout the specifications. Furthermore, both the Q and τ2 statistics indicated substantial heterogeneity.
Table 5 reports whether the coefficients (marginal associations) are significantly different from zero, but not whether they are significantly different from each other. In Table 6, we report results where we used the most comprehensive specification (column 2) of Table 5 to examine if coefficients are significantly different from each other (see the note below the table for details on how we implemented the test). To keep the table at a manageable length, we focused on the three components with statistically significant coefficients that were stable across specifications: peer‐assisted instruction, small‐group instruction, and fractions. We compared peer‐assisted instruction and small‐group instruction to the other instructional methods and fractions with the other math domains.
Peer‐assisted instruction and small‐group instruction were not significantly different from one another. Peer‐assisted instruction was associated with significantly larger effect sizes than all other instructional methods. Small‐group instruction was associated with significantly larger effect sizes than all but one instructional method, medium‐group instruction where p = .07. Fractions were associated with significantly larger effect sizes than all other math domains. Note that we are testing multiple hypotheses here and that the reported p values are not adjusted for the number of tests. As we did not know the type of tests and number of hypotheses to be tested beforehand, our protocol did not specify an adjustment procedure.
The multiple meta‐regressions provide an estimate of the isolated association between each component and effect sizes, conditional on other components and study characteristics. As discussed in Section, it is difficult to rule out that we did not introduce bias by including moderators in the regressions. Furthermore, we were unable to include interactions between components in these regressions. The reason is that there were few recurring combinations in our sample. In the next section, we therefore used subgroup analysis to examine the effect sizes of these recurring combinations. 4.3.10
| (1) | (2) | (3) | (4) | |||||
|---|---|---|---|---|---|---|---|---|
| Moderator | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI |
| CAI | 0.1 | [−0.02,0.21] | 0.08 | [−0.07, 0.23] | 0.17 | [−0.02, 0.36] | −0.14 | [−0.28, 0.01] |
| Coaching | 0.01 | [−0.11,0.12] | −0.03 | [−0.15, 0.09] | 0.03 | [−0.12, 0.17] | −0.23 | [−0.40, −0.06] |
| Incentives | 0.08 | [−0.04,0.21] | 0.09 | [−0.06, 0.24] | 0.06 | [−0.08, 0.20] | 0.07 | [−0.08, 0.23] |
| Medium‐group | 0.14 | [−0.06,0.34] | 0.18 | [−0.10, 0.47] | 0.28 | [−0.07, 0.63] | −0.34 | [−0.78, 0.11] |
| Other method | 0.1 | [−0.07,0.27] | ||||||
| Peer‐assisted | 0.41 | [0.24,0.59] | 0.4 | [0.13, 0.67] | 0.42 | [0.07, 0.76] | 0.38 | [0.12, 0.65] |
| .Progress mon | −0.12 | [−0.22,‐0.01] | −0.09 | [−0.20, 0.02] | −0.09 | [−0.21, 0.03] | −0.07 | [−0.27, 0.13] |
| Small‐group | 0.36 | [0.26,0.47] | 0.34 | [0.12, 0.56] | 0.32 | [0.07, 0.58] | 0.24 | [0.11, 0.37] |
| Comprehension | 0 | [−0.15,0.14] | 0.04 | [−0.10, 0.18] | ||||
| Decoding | 0.04 | [−0.07,0.15] | 0.01 | [−0.11, 0.13] | ||||
| Fluency | −0.06 | [−0.20,0.08] | −0.07 | [−0.20, 0.06] | ||||
| Multiple reading | 0 | [−0.15,0.14] | ||||||
| Spelling/writing | 0.11 | [−0.02,0.23] | 0.11 | [0.00, 0.23] | ||||
| Vocabulary | −0.08 | [−0.19,0.03] | −0.07 | [−0.19, 0.05] | ||||
| Algebra | −0.13 | [−0.27,0.02] | −0.16 | [−0.31, −0.01] | ||||
| Fractions | 0.33 | [0.01,0.65] | 0.37 | [0.06, 0.67] | ||||
| Geometry | 0.03 | [−0.10,0.15] | 0.04 | [−0.15, 0.22] | ||||
| Multiple math | −0.01 | [−0.23,0.21] | ||||||
| Number sense | 0.09 | [−0.09,0.27] | −0.02 | [−0.20, 0.16] | ||||
| Operations | −0.08 | [−0.26,0.10] | −0.10 | [−0.27, 0.07] | ||||
| Problem solving | 0.05 | [−0.07,0.17] | 0.08 | [−0.04, 0.19] | ||||
| Gen. academic | −0.05 | [−0.24,0.15] | −0.09 | [−0.33, 0.15] | ||||
| Meta‐cognitive | −0.05 | [−0.20,0.10] | 0 | [−0.14, 0.13] | ||||
| Social‐emotional | −0.07 | [−0.38,0.25] | −0.07 | [−0.39, 0.24] | ||||
| QES | −0.04 | [−0.22, 0.15] | 0.04 | [−0.14, 0.21] | 0.03 | [−0.40, 0.46] | ||
| General test | 0.07 | [−0.10, 0.25] | 0.07 | [−0.13, 0.26] | −0.03 | [−0.19, 0.12] | ||
| Mean Grade | −0.03 | [−0.06, −0.00] | −0.04 | [−0.07,0.02]− | −0.01 | [−0.05, 0.02] | ||
| Math | 0.07 | [−0.12, 0.26] | ||||||
| Constant | 0 | [−0.21, 0.20] | −0.03 | [−0.31, 0.24] | 0.19 | [0.01, 0.37] | ||
| Effect sizes | 1030 | 1030 | 829 | 199 | ||||
| Study clusters | 189 | 189 | 138 | 64 | ||||
| N | 206,186 | 206,186 | 132,046 | 74,608 | ||||
| Q | 481 | 430.8 | 343.3 | 142.6 | ||||
| I2 | 65.7 | 62.2 | 63 | 62.8 | ||||
| τ2 | 0.063 | 0.053 | 0.051 | 0.047 | ||||
Specific combinations of instructional methods
We found few recurring combinations of instructional methods in our data. Only two analyses of pairs of instructional methods produced adjusted degrees of freedom above 4: coaching of personnel and small‐group instruction, and incentives and small‐group instruction. The first combination, examined in nine studies, had a lower average effect size (ES = 0.31, CI = [0.10, 0.52]) than small‐group instruction alone (compare Table 4). The incentives and small‐group instruction combination, examined in seven studies, had a larger effect size than small‐group instruction alone (ES = 0.47, CI = [0.36, 0.58]). We found no combination of three or more instructional methods examined in more than two studies.
Peer‐assisted instruction and small‐group instruction
The only two instructional methods with stable, large, and statistically significant associations with effect sizes in the previous analyses were peer‐assisted instruction and small‐group instruction. Our definitions of the peer‐assisted and small‐group instruction categories are relatively broad and they may contain diverse interventions. We therefore examined them further in this section. We focused on single method interventions and short‐term effects to better isolate the contribution of the instructional method and reduce heterogeneity due to measurement timing.
Most other instructional methods have been studied in few single method interventions and were part of few recurring combinations. Partial exceptions were CAI (examined in 12 studies of single method interventions) and medium‐group instruction (examined in 10 studies of single method interventions). The multiple meta‐regressions indicated that both these methods had different associations with effect sizes based on math and reading tests. However, only 4 out of 12 CAI studies and only 1 out of 10 medium‐group studies tested effects using math tests.Examining this issue further was therefore difficult. All content domains that the previous meta‐regressions indicated were associated with larger effect sizes have been examined in few single domain interventions. 11
The peer‐assisted instruction category is less diverse than it may seem from our definition. Only 3 out of 16 single method studies used cross‐age peer‐tutoring, the rest was interventions where same‐age peers worked together (often called cooperative learning). The average effect size for the 13 same‐age peers studies was slightly smaller than the one reported in Table 4 for peer‐assisted instruction, but still large and statistically significant (ES = 0.32, CI = [0.25, 0.42]). The average effect size in the three cross‐age peer‐tutoring studies was large (ES = 0.85, CI = [−0.36, 2.1]) but the adjusted degrees of freedom was below 4 and the result was unreliable. Four studies used larger peer‐groups than pairs. The effect size was larger when pairs were used (ES = 0.42, CI = [0.27, 0.58]) than when larger groups were used (ES = 0.28, CI = [−0.22, 0.78]). The effect size in the 6 peer‐assisted instruction studies using math tests was larger (ES = 0.54, CI = [0.06, 1.0]) than the effect size in the 13 studies testing reading (ES = 0.33, CI = [0.22, 0.44]), but the degrees of freedom were only 4.5 in the analysis of math tests so the result should be viewed with caution. The uniformity of peer‐assisted interventions may be one explanation of the relatively low level of heterogeneity reported in Table 4. Furthermore, the prediction interval based on the analysis Table 4 does not include zero, but ranges from 0.06 to 0.71.
Adult‐led small‐group instruction is a more diverse and larger category of interventions than peer‐assisted instruction (there were 85 studies of single method small‐group instruction interventions). The heterogeneity reported in Table 4 is also substantial, and a prediction interval based on this analysis ranges from −0.20 to 0.95. Interventions in this group of studies targeted either subjects like reading and mathematics or non‐subject‐specific areas like social‐emotional skills. Although it is difficult to draw a sharp line between the two, the former are usually some form of tutoring while the latter are more often called mentoring. However, if we define mentoring as interventions that do not target any subject‐specific domain, our sample only includes three studies of such interventions. Thus, with these definitions of tutoring and mentoring, our results for small‐group instruction mainly pertains to tutoring interventions. The effect sizes in single method small‐group interventions were reasonably similar in math (37 studies, ES = 0.39, CI = [0.30, 0.48]) and reading (65 studies, ES = 0.34, CI = [0.26, 0.42]).
We therefore combined the subjects and focused the further examination on group sizes and additional study characteristics. Medium‐group instruction is only different from small‐group instruction by our definition of small and medium, and we therefore included single medium‐group instruction interventions in this examination as well to increase statistical power (10 studies). In total, this left us with 95 studies, 157 interventions, and 583 effect sizes.
In column 1 of Table 7, we split up the small‐group instruction group into three: instruction one‐to‐one, one‐to‐two or three, and one‐to‐four or five. The reference category in this specification, and in column 2 of this table, is the medium‐group instruction category, where groups range from 6 to 20 (or are of unclear size). There were 67 interventions of one‐to‐one instruction, 48 of one‐to‐two or three, 35 of one‐to‐four or five, and 11 in the medium‐group category. Group sizes sometimes vary between interventions within studies and some interventions used more than one group size (e.g., both one‐to‐one and one‐to‐two), so the sum count to more than 157, although no intervention use both small‐group and medium‐group instruction. The small‐group instruction categories are all associated with larger effect sizes than the medium‐group category, although the differences are not significant. The coefficients on the one‐to‐one, one‐to‐two or three, and one‐to‐four or five are close to one another.
Column 2 adds the study characteristics without missing values, which for these interventions include duration. Adding study characteristics make the differences between the small‐group categories and the medium‐group category even smaller. Only the mean Grade is statistically significant among the study characteristics. As in the full sample, the coefficient is negative and indicates that interventions in higher Grades are associated with smaller effects. Note further that the I2, τ2, and Q‐statistics indicate that there was still systematic variation and substantial heterogeneity in this restricted sample.
Although this analysis tried to isolate associations between group size and effect sizes, the regressions are unlikely to uncover the causal effect of group size reductions. For example, one‐to‐one tutoring may be used for students with the greatest academic difficulties whereas larger groups are used for students with less grave difficulties. As mentioned, we were unable to control directly for the students’ level of difficulties. Therefore, if it is harder to improve the achievement of the group with the greatest difficulties or if group sizes work differently depending on the level of difficulties, then the group size associations are confounded by the students’ level of difficulties.
A better way to examine the effects of group size is to use comparison designs and meta‐analyse interventions that changed the group size, while keeping everything else constant. That is, studies that assign the group size randomly or quasi‐experimentally. However, we found only four studies that contrasted a one‐to‐one tutoring programme with the same, or a highly similar, programme using groups of two to five students (three RCTs and one QES). Running a meta‐analysis on these four studies yielded a negative (indicating an advantage of one‐to‐one tutoring) but far from significant effect size (ES = −0.14, CI = [−0.70, 0.42], 4 studies, 24 effect sizes, 658 student observations). Moreover, the adjusted degrees of freedom was below 4. While the heterogeneity statistics did not indicate significant heterogeneity, the number of studies likely imply that the estimations and the statistics are unreliable and the test of heterogeneity underpowered.
As mentioned, RVE may have trouble estimating the heterogeneity and standard errors when the number of studies is very small. However, the conclusion about group sizes was not sensitive to changes of the specification. We used study‐level averages to estimate the between‐study variance (using the REML option in the R package metafor), adjusted for pretest differences in the one study for which we based the effect sizes on raw means, and excluded the one intervention that used groups of five students (all others contrasted one‐to‐one with groups of two or three). The results were very similar (see Supporting Information Appendix K: Extra sensitivity analyses).
Results of sensitivity analyses
The sections below report results from our sensitivity analyses. We focused the sensitivity analyses on our main results: we found positive, substantial, and statistically significant overall average short‐term and follow‐up effect sizes, and that peer‐assisted instruction and small‐group instruction had large, stable, and statistically significant average effect sizes across specifications. For all other instructional methods and all content domains, our results were either based on few studies, particularly of single method or single domain interventions, or were not stable across specifications. We are therefore hesitant to make conclusions about their effectiveness and, as finding that the results are sensitive would not change any conclusions, we did not run the sensitivity tests for them.
We tested whether effect sizes were associated with effect size measurement, whether they were sensitive to adjusting for outliers and different adjustment for clustered assignment of treatment, to multiple imputation of moderators with missing values, and to adjusting for the risk of bias ratings. We also examined whether there were signs of heterogeneity across control group conditions, and finally, if there were indications of publication bias.
We present some of the results of these sensitivity analyses in four figures (Figures 10, 11, 12, 13), corresponding to overall short‐term effects, overall follow‐up effects, peer‐assisted instruction, and small‐group instruction. Each figure has the effect size and its CI from the primary analysis at the bottom of the figure for easy reference. As we believe the single method interventions have the best chances of isolating the effects of instructional methods, we used them in the analyses of peer‐assisted instruction and small‐group instruction. We confined the sensitivity analysis to short‐term effects for peer‐assisted and small‐group instruction, but recall that the follow‐up effects were largely from small‐group instruction interventions.
Table 8 reports the results from the sensitivity analyses of effect size measurement, outliers, clustered assignment of treatment, and risk of bias using the most comprehensive specification among our meta‐regressions (reported in column 2 of Table 5). As these meta‐regressions are the only specifications including moderators in the primary analysis, we conducted the multiple imputation of moderators using this specification. Similarly, we examined the heterogeneity of control group progression using this specification. We report multiple meta‐regressions from the latter two types of analyses in Table 9. We comment on the figures and table by type of sensitivity analysis below.
Sensitivity analyses: Overall short‐term average effect size
Sensitivity analyses: Overall follow‐up average effect size
Sensitivity analyses: Peer‐assisted instruction
Sensitivity analyses: Small‐group instruction
| ES measurement | Outliers | Clustered | Risk of bias | |||||
|---|---|---|---|---|---|---|---|---|
| Moderator | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI |
| CAI | 0.09 | [−0.07, 0.25] | 0.07 | [−0.06, 0.20] | 0.07 | [−0.08, 0.22] | 0.07 | [−0.11, 0.25] |
| Coaching | −0.02 | [−0.14, 0.10] | −0.02 | [−0.14, 0.09] | −0.03 | [−0.15, 0.08] | −0.02 | [−0.15, 0.10] |
| Incentives | 0.09 | [−0.06, 0.25] | 0.08 | [−0.06, 0.22] | 0.09 | [−0.06, 0.23] | 0.08 | [−0.09, 0.26] |
| Medium‐group | 0.18 | [−0.11, 0.47] | 0.15 | [−0.07, 0.38] | 0.17 | [−0.09, 0.43] | 0.21 | [−0.20, 0.62] |
| Peer‐assisted | 0.4 | [0.12, 0.69] | 0.36 | [0.14, 0.59] | 0.39 | [0.13, 0.64] | 0.39 | [0.04, 0.73] |
| .Progress mon | −0.09 | [−0.20, 0.03] | −0.08 | [−0.19, 0.02] | −0.08 | [−0.19, 0.02] | −0.09 | −[0.22, 0.03] |
| Small‐group | 0.34 | [0.11, 0.57] | 0.31 | [0.14, 0.48] | 0.32 | [0.11, 0.53] | 0.33 | [0.06, 0.59] |
| Comprehension | 0.03 | [−0.12, 0.17] | 0.04 | [−0.09, 0.17] | 0.04 | [−0.09, 0.18] | 0.01 | [−0.15, 0.16] |
| Decoding | 0 | [−0.12, 0.13] | 0.01 | [−0.10, 0.13] | 0 | [−0.12, 0.12] | 0.05 | [−0.09, 0.20] |
| Fluency | −0.05 | [−0.19, 0.08] | −0.06 | [−0.19, 0.06] | −0.06 | [−0.19, 0.06] | −0.06 | [−0.21, 0.10] |
| Spelling/writing | 0.11 | [−0.01, 0.22] | 0.11 | [0.00, 0.22] | 0.12 | [0.01, 0.24] | 0.09 | [−0.03, 0.22] |
| Vocabulary | −0.07 | [−0.19, 0.05] | −0.07 | [−0.18, 0.04] | −0.06 | [−0.17, 0.05] | −0.08 | [−0.20, 0.05] |
| Algebra | −0.16 | [−0.31, −0.01] | −0.16 | [−0.31, −0.01] | −0.15 | [−0.30, −0.01] | −0.17 | [−0.35, 0.02] |
| Fractions | 0.37 | [0.06, 0.67] | 0.36 | [0.05, 0.67] | 0.37 | [0.09, 0.65] | 0.33 | [0.05, 0.62] |
| Geometry | 0.03 | [−0.17, 0.22] | 0.02 | [−0.15, 0.18] | 0.03 | [−0.15, 0.22] | 0.04 | [−0.20, 0.28] |
| Number sense | −0.01 | [−0.19, 0.18] | −0.02 | [−0.19, 0.16] | −0.01 | [−0.18, 0.16] | −0.09 | [−0.27, 0.10] |
| Operations | −0.11 | [−0.28, 0.06] | −0.10 | [−0.26, 0.06] | −0.10 | [−0.26, 0.06] | −0.09 | [−0.27, 0.09] |
| Prob. solving | 0.07 | [−0.04, 0.19] | 0.08 | [−0.04, 0.19] | 0.07 | [−0.04, 0.19] | 0.1 | [−0.04, 0.24] |
| Gen. academic | −0.08 | [−0.32, 0.16] | −0.09 | [−0.28, 0.11] | −0.06 | [−0.28, 0.16] | −0.12 | [−0.36, 0.12] |
| Meta‐cognitive | −0.02 | [−0.17, 0.13] | 0.01 | [−0.11, 0.13] | 0 | [−0.14, 0.14] | 0.01 | [−0.13, 0.16] |
| Social‐emotional | −0.08 | [−0.41, 0.25] | −0.10 | [−0.36, 0.16] | −0.09 | [−0.40, 0.23] | −0.09 | [−0.43, 0.24] |
| QES | −0.04 | [−0.22, 0.15] | −0.02 | [−0.19, 0.15] | −0.05 | [−0.24, 0.15] | ||
| General test | 0.06 | [−0.11, 0.24] | 0.06 | [−0.09, 0.20] | 0.07 | [−0.10, 0.24] | 0.08 | [−0.12, 0.27] |
| Mean Grade | −0.03 | [−0.06, ‐0.00] | −0.03 | [−0.06, −0.01] | −0.03 | [−0.06, −0.01] | −0.03 | [−0.05, 0.00] |
| Math | 0.07 | [−0.12, 0.26] | 0.08 | [−0.08, 0.24] | 0.07 | [−0.12, 0.26] | 0.11 | [−0.10, 0.33] |
| Raw mean | 0.03 | [−0.06, 0.13] | ||||||
| Glass's δ | 0.24 | [−0.01, 0.48] | ||||||
| Unclear ES type | −0.16 | [−1.11, 0.80] | ||||||
| .Super‐pop | 0.01 | [−0.20, 0.22] | ||||||
| Blinding | 0.01 | [−0.08, 0.10] | ||||||
| .Incomplete out | −0.04 | [−0.17, 0.09] | ||||||
| Reporting | −0.05 | [−0.16, 0.07] | ||||||
| Other bias | 0.02 | [−0.09, 0.13] | ||||||
| Constant | −0.02 | [−0.24, 0.21] | 0.01 | [−0.16, 0.18] | 0.01 | [−0.19, 0.20] | 0 | [−0.26, 0.26] |
| Effect sizes | 1030 | 1030 | 1030 | 981 | ||||
| Study clusters | 189 | 189 | 189 | 172 | ||||
| N | 206,186 | 206,186 | 206,186 | 152,581 | ||||
| Q | 412.1 | 382.3 | 488 | 385.3 | ||||
| I2 | 61.4 | 57.4 | 66.6 | 62.9 | ||||
| τ2 | 0.056 | 0.043 | 0.048 | 0.058 | ||||
| Multiple imputation | Control progression | Publishing status | Asymmetry | |||||
|---|---|---|---|---|---|---|---|---|
| Moderator | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI | Coef. | 95% CI |
| CAI | 0.08 | [−0.07, 0.21] | −0.08 | [−0.39, 0.23] | 0.07 | [−0.07, 0.21] | 0.05 | [−0.09, 0.19] |
| Coaching | −0.03 | [−0.15, 0.10] | 0.62 | [−0.22, 1.46] | −0.02 | [−0.15, 0.10] | 0.04 | [−0.09, 0.16] |
| Incentives | 0.09 | [−0.06, 0.24] | −0.23 | [−0.62, 0.16] | 0.09 | [−0.06, 0.24] | 0.1 | [−0.05, 0.24] |
| .Medium‐gr | 0.18 | [−0.10, 0.43] | 0.07 | [−0.27, 0.40] | 0.19 | [−0.10, 0.48] | 0.17 | [−0.10, 0.44] |
| .Peer‐assist | 0.38 | [0.13, 0.61] | −0.02 | [−0.25, 0.28] | 0.4 | [0.13, 0.66] | 0.4 | [0.13, 0.68] |
| Progress | −0.08 | [−0.19, 0.04] | 0.2 | [−0.15, 0.55] | −0.10 | [−0.21, 0.02] | −0.08 | [−0.20, 0.04] |
| Small‐group | 0.34 | [0.12, 0.54] | 0.17 | [−0.12, 0.45] | 0.34 | [−0.12, 0.55] | 0.3 | [0.10, 0.50] |
| .Comprehen | 0.04 | [−0.10, 0.18] | −0.30 | [−0.62, 0.02] | 0.04 | [−0.10, 0.18] | 0.04 | [−0.10, 0.17] |
| Decoding | 0.01 | [−0.13, 0.11] | −0.09 | [−0.52, 0.33] | 0.02 | [−0.10, 0.14] | −0.01 | [−0.13, 0.11] |
| Fluency | −0.07 | [−0.18, 0.07] | 0.12 | [−0.14, 0.38] | −0.07 | [−0.20, 0.06] | −0.07 | [−0.20, 0.05] |
| Spell./writing | 0.11 | [0.00, 0.24] | 0.05 | [−0.22, 0.32] | −0.12 | [0.00, 0.23] | 0.1 | [−0.01, 0.22] |
| Vocabulary | −0.06 | [−0.16, 0.07] | 0.09 | [−0.24, 0.42] | −0.08 | [−0.19, 0.04] | −0.06 | [−0.18, 0.05] |
| Algebra | −0.16 | [−0.33, −0.01] | −0.66 | [−1.48, 0.16] | −0.16 | [−0.31, −0.01] | −0.17 | [−0.33, −0.02] |
| Fractions | 0.36 | [0.08, 0.62] | 0.07 | [−0.75, 0.90] | 0.37 | [0.07, 0.68] | 0.38 | [−0.08, 0.67] |
| Geometry | 0.06 | [−0.11, 0.25] | −0.11 | [−0.82, 0.59] | 0.04 | [−0.15, 0.22] | 0.04 | [−0.16, 0.24] |
| .Number sen | −0.02 | [−0.19, 0.15] | 0.57 | [0.10, 1.03] | −0.02 | [−0.21, 0,17] | −0.03 | [−0.20, 0.15] |
| Operations | −0.10 | [−0.29, 0.05] | −0.05 | [−0.65, 0.55] | −0.09 | [−0.27, 0,09] | −0.09 | [−0.26, 0.07] |
| Prob. solving | 0.07 | [−0.04, 0.19] | 0.18 | [−0.50, 0.87] | 0.08 | [−0.04, 0.21] | 0.08 | [−0.03, 0.19] |
| .Gen. acad | −0.09 | [−0.30, 0.17] | 0.13 | [−0.60, 0.86] | −0.08 | [−0.31, 0.16] | −0.09 | [−0.34, 0.15] |
| .Meta‐cog | −0.01 | [−0.15, 0.12] | 0.02 | [−0.27, 0.31] | 0 | [−0.14, 0.13] | −0.02 | [−0.16, 0.12] |
| .Social‐emot | −0.07 | [−0.42, 0.23] | −0.29 | [−0.61, 0.03] | −0.07 | [−0.39, 0.24] | −0.08 | [−0.39, 0.23] |
| QES | −0.04 | [−0.25, 0.15] | 0.13 | [−0.29, 0.55] | −0.05 | [−0.25, 0.15] | −0.02 | [−0.20, 0.15] |
| General test | 0.08 | [−0.09, 0.25] | 0.03 | [−0.43, 0.49] | 0.06 | [−0.11, 0.23] | 0.08 | [−0.10, 0.25] |
| Mean Grade | −0.03 | [−0.06, 0.00] | −0.10 | [−0.17, −0.03] | −0.03 | [−0.06, −0.01] | −0.03 | [−0.06, −0.01] |
| Math | 0.06 | [−0.12, 0.27] | −0.05 | [−0.41, 0.30] | 0.07 | [−0.12, 0.26] | 0.06 | [−0.13, 0.25] |
| .Imp. prob | −0.04 | [−0.14, 0.08] | ||||||
| Share girls | −0.00 | [−0.01, 0.00] | ||||||
| Minority | 0 | [−0.00, 0.00] | ||||||
| Duration | 0 | [−0.01, 0.00] | ||||||
| School staff | 0.03 | [−0.10, 0.11] | ||||||
| Journal | −0.06 | [−0.21, 0.08] | ||||||
| .Large pop | −0.05 | [−0.14, 0.04] | ||||||
| Clustered | −0.11 | [−0.22, −0.01] | ||||||
| Constant | 0.09 | [−0.27, 0.45] | 0.51 | [−0.04, 1.07] | 0.05 | [−0.15, 0.25] | 0.1 | [−0.10, 0.29] |
| Effect sizes | 1030 | 637 | 1030 | 1030 | ||||
| Clusters | 189 | 142 | 189 | 189 | ||||
| N | 201,734 | 120,925 | 206,186 | 206,186 | ||||
| Q | 412.2 | 2393.2 | 430.3 | 425.2 | ||||
| I2 | 61.7 | 95.2 | 62.4 | 62.1 | ||||
| τ2 | 0.055 | 0.714 | 0.054 | 0.054 | ||||
Effect size measurement
To test sensitivity to how effect sizes were measured and calculated, we included four moderators indicating whether we used the raw means to calculate the effect size, standardised the SMDs with the control group standard deviation (i.e., used a Glass's δ), if the standardisation was unclear, or the effect size was standardised with a standard deviation from a super‐population. There were no Glass's δ, unclear effect size types, and effect sizes standardised with a super‐population among the follow‐up outcomes and among the single method peer‐assisted instruction effect sizes. In these regressions, we just included the raw means‐indicator. There were no Glass's δ and no effect sizes standardised with a super‐population among the single method small‐instruction effect sizes. Thus, this regression included the raw means‐indicator and the indicator for unclear effect sizes.
Including effect size measurement‐moderators in the analyses did not change the estimated average effect sizes much (all increased somewhat). Although the average effect sizes retained their statistical significance in all four analyses, the CIs became broader in all cases. This was expected, as many effect sizes in all analyses was measured or calculated differently and the effect size shown in the figures is the average among those that were not. In particular, we used the raw means to calculate a relatively large proportion of the effect sizes.
The raw means‐indicator was not statistically significant in any analysis, and changed sign between specifications. Effect sizes of unclear type were associated with smaller effect sizes in the three specifications in which we could include it and so were those standardised with a super‐population in the analysis of overall short‐term effects. The only statistically significant moderator was the unclear effect size type in the analysis of small‐group instruction interventions (β = −.27, CI = [−0.40, −0.14]).
In the multiple meta‐regression in Table 8, Glass's δ and unclear effect size type are relatively large but not significant, whereas raw means and standardising with a super‐population are small and not significant. Peer‐assisted instruction and small‐group instruction retained both their magnitude and statistical significance in the meta‐regression compared with the primary analysis.
Outliers
We examined the distributions of effect sizes for the presence of outliers and the sensitivity of our main results by methods suggested by Lipsey and Wilson (2001): trimming the distribution by dropping the outliers and by winsorizing the outliers to the nearest non‐outlier value. We show the latter results in the figures below. Supporting Information Appendix K: Extra sensitivity analyses contains figures showing the effect size distributions of the four types of effect sizes and the results of additional analyses.
Although is difficult to come up with a definition of outliers that is not in some sense arbitrary, there are quite a few effect sizes, particularly in the small‐group instruction category and among the short‐term effects (many of which are the same), which seem like clear outliers. Outliers seem rarer among the follow‐up and peer‐assisted effect sizes, although there a few potential examples. The short‐term and small‐group distributions start to thin out around 1.5 and −0.5, respectively. We used these values as cut‐offs when winsorizing.
The studies with larger and smaller effect sizes than these cut‐offs were almost all small sample studies. All except one effect size was based on a total sample size of 60 or under. Otherwise, they were not exceptional in terms of other potential explanations, for example, study design, risk of bias, or whether we used raw means to calculate effect sizes or not.
Winsorizing outliers had a small impact on our results. All average effect sizes in the subgroup analyses are reasonably close to those in the primary analysis and were still statistically significant (the follow‐up effect size increased to 0.34, which is the largest difference). Peer‐assisted instruction and small‐group instruction were still sizeable and statistically significant in the multiple meta‐regression.
In Supporting Information Appendix K, we report results from analyses where we sequentially remove outliers down to effect sizes in between −0.25 and 1. The results are also again relatively close to those in the primary analysis. Thus, outliers do not seem to be driving our results. It is also worth noting that when we removed outliers, the heterogeneity decreased by quite a lot. For example, with the harshest cut‐off, the I2 was 34.4% and the τ2 was 0.02 in the meta‐regression including both instructional methods, content domains, and study characteristics.
Clustered assignment of treatment
We tested sensitivity to clustered assignment of treatment by the methods described in Section 4.3.5. In the effect size estimates shown in Figures 10, 11, 12, 13 (named “4. Clustered”), we did not adjust for clustering. We report results in text from specifications in which we instead adjusted effect sizes using a substantially higher ICC (0.3) than in the primary analysis.
As can be seen in Figures 10, 11, 12, 13 and for peer‐assisted instruction and small‐group instruction in Table 8, not adjusting for clustered assignment of treatment left our estimates virtually unchanged. Although adjusting for clustering decreases the individual effect sizes, the average effect sizes decreased slightly when we used unadjusted effect sizes. The reason is likely connected to the fact that the variances of studies with clustered assignment treatment is larger when we adjust for clustering. Therefore, adjusting for clustering gives the adjusted effect sizes smaller weights in the meta‐analyses. As effect sizes from studies using clustered assignment of treatment tend to be smaller in our sample, the average effect size decreased when these studies receive more weight.
A more surprising result was that the CI actually increased a little with unadjusted effect sizes in the analysis of follow‐up effects (the other intervals became slightly broader). The reason may be that the between‐study heterogeneity decreased as well (τ2 increased from 0.031 to 0.042). If effect sizes using clustered assignment were further away from the average than effect sizes using individual assignment when not adjusted, they will receive less weight when we adjust. If the ensuing reduction in τ2 is large enough, it will dominate the effect of increasing the individual variances by adjusting for clustering.
Increasing the ICC to 0.3 strengthened the above tendencies, although the differences to the primary analysis were again very small. All our main results retained their significance also with this substantially higher ICC (see Supporting Information Appendixfor these results). K
Risk of bias
We used the items with numerical ratings from the risk‐of‐bias assessment to examine if the ratings were associated with effect sizes. As described in Section, we coded indicator variables that contrasted effect sizes given a higher risk of bias rating to effect sizes with a lower risk. We defined higher risk as 4 or unclear for the items blinding, incomplete outcome reporting, and other bias, and as ratings higher than 1 for the selective outcome reporting item. We coded all indicators so that they equalled 1 for ratings indicating a higher risk of bias and included them in the multiple meta‐regressions. 4.3.11
As discussed in Section, we omitted the QES from this sensitivity analysis. The coefficient on the constant in the meta‐regressions for overall short‐term effects, overall follow‐up effects, peer‐assisted instruction, and small‐group instruction should therefore be interpreted as the average effect size in RCTs with a relatively low risk of bias. In the analysis of peer‐assisted instruction, we could not include the blinding indicator because there were only two studies rated low risk. Including it together with the other indicators meant that the reference category would have been empty. The constant in the meta‐regressions including both intervention components and study characteristics without missing values is the average effect size in RCTs that did not use any of the included components/characteristics and was at the mean of the mean‐centred variables. 4.3.11
All average effect sizes in Figures 10, 11, 12, 13 increased slightly when we adjusted for the risk of bias ratings compared with the primary analysis. That is, effect sizes with less risk of bias tended to be slightly larger, although the differences were minor. The CIs became broader in all analyses, reflecting that we rated relatively few studies as having a (relatively) low risk of bias on all included items. However, all average effect sizes in RCTs with a low risk of bias were statistically significantly different from zero. Both peer‐assisted instruction and small‐group instruction retained both the magnitude and statistical significance in the multiple meta‐regression. The risk of bias‐item indicators all had small and statistically insignificant associations with effect sizes in this regression.
Moderatorwith missing values s
In the multiple meta‐regressions reported in Table 5, we omitted potentially important moderators because they had missing values. In this section, we report results from specifications using multiple imputation to account for missing values of moderators with relatively low rate of missing values, as described in Section 4.3.6. We imputed values for the following moderators: an indicator for implementation problems, share of girls, share of minority students, duration, and an indicator for whether school staff implemented the intervention (as opposed to e.g., researchers or volunteers). Information about these moderators were missing from <20% of interventions and they are potentially relevant for all types of interventions. Moderators such as the number of sessions and hours per week were also missing from <20% of interventions but are not relevant for all types of interventions (e.g., there are no sessions in incentive and progress monitoring interventions).
The results in column 1 of Table 9 indicate that adjusting for these moderators does not change our main results. The coefficients on peer‐assisted instruction and small‐group instruction are of similar size and both are still statistically significant. Furthermore, all other coefficients are close to the values displayed in column 2, Table 5. The coefficients on the imputed moderators are all small and not statistically significant.
Heterogeneity of controgroup progression l
This section examines whether there is heterogeneity across the control group conditions by examining the control group's progression from pre‐ to posttest. As in our analysis of heterogeneity of the overall effect sizes, we confined this analysis to the short‐term effects. We calculated a control group “effect size”, as the difference between pre‐ and posttest divided by the control group posttest standard deviation. We used the posttest standard deviation to be able to better compare the progress with the effect sizes found in the primary analysis.
We then (a) tested whether this control group “effect size” was heterogeneous across studies and (b) used multiple meta‐regression to examine whether study characteristics explained some of the heterogeneity. We included 142 studies that supplied the necessary information in the analysis. That is, this sample is different from our primary analysis sample and the results reported should be viewed with some caution in relation to the heterogeneity of control group progression in the primary analysis sample.
We found substantial heterogeneity in the control group progression across studies. We first estimated a specification including just a constant using the same RVE procedure as in the primary analysis (not shown in Table 9). The constant therefore shows the weighted average control group progression, which is 0.62 (CI = [0.50, 0.74]). There is substantial variation around this average: the τ2 is 0.70, the I2 is 96%, and the Q‐statistic is 3949, which is highly significant. There continues to be heterogeneity when we confine the set of interventions to those using peer‐assisted instruction (mean = 0.47, τ2 = 0.07, I2 = 55.4, and Q = 31.4) or small‐group instruction (mean = 0.61, τ2 = 0.28, I2= 88.2, and Q = 534.3) as their only instructional method.
The results in column 2 of Table 9, which mimics the specification shown in Table 5, column 2, shows that very few of the instructional methods and content domains were significantly associated with the control group progression. The only two significant associations are with number sense and mean Grade. Number sense is positively associated and mean Grade negatively associated with effect sizes. These results are not surprising, given that students on average make less progress, the higher the Grade (Lipsey et al., 2012), and that number sense is typically a focus in the early years of primary school.
Thus, the results indicated that control group progression was strongly heterogeneous, which may imply that the quality of control group instruction is an important explanation of the heterogeneity we see in most of our analyses. However, it was reassuring that we did not find strong evidence of an association between the control group progression and intervention components, and in particular that we did not find a significant association with peer‐assisted instruction and small‐group instruction.
Publication bias
This section examines publication bias by testing whether unpublished studies have different effect sizes compared with published studies, and by using funnel plots and Egger's test (Egger et al., 1997). However, we want to acknowledge that these tests are for several reasons difficult to interpret as direct evidence of publication bias, or the lack thereof. The effect size estimates and corresponding standard errors we have analysed are not necessarily the ones used by authors, editors, and reviewers to decide whether a paper should be published. We used only effect sizes from standardised tests that had low enough risk of bias, which were included in studies found through our search and screening process. There may thus be publication bias in the literature that would not show up in our sample and other processes than publication bias can cause asymmetries in funnel plots. We discuss the interpretation of funnel plots further below.
We found no evidence that effect sizes from studies published in scientific journals were larger than effect sizes from studies not published in journals (e.g., in government reports, working papers, and dissertations). We added an indicator equal to 1 if the study was published in a journal to the specification in column 2 of Table 5 for the full sample of short‐term effects. The estimate, reported in column 3 of Table 9, indicated that studies published in journals, conditional on all other moderators without missing values, have slightly lower effect sizes (β = −.06), but not significantly so.
For the funnel plots and test of asymmetry, we averaged effect sizes and variances over studies and estimated a random‐effects model by using the REML option with the Knapp and Hartung adjustment of standard errors in the R package metafor (Viechtbauer, 2010). As mentioned, this procedure also provided a sensitivity test of our primary analysis.
Table 10 shows the average effect sizes, CIs, and heterogeneity statistics for the REML procedure compared with the RVE procedure used in the primary analysis. The effect sizes and CIs from the two procedures are very close in all four cases. The heterogeneity statistics indicate more heterogeneity in the RVE procedure, which seems reasonable as the RVE procedure takes into account variation in effect sizes within studies. However, our conclusions would be similar using study‐level averages. There is substantial heterogeneity across short‐term and small‐group instruction effect sizes. The heterogeneity across follow‐up effect sizes is smaller, but still statistically significant, whereas the heterogeneity among the effect sizes from peer‐assisted instruction intervention is relatively small and not significant.
Figure 13 displays funnel plots of the study‐level short‐term effect sizes (upper left corner), follow‐up effect sizes (upper right corner), peer‐assisted instruction effect sizes (lower left corner), and small‐group instruction effect sizes (lower right corner). The effect sizes are shown on the x‐axis and standard errors on the y‐axis. The center‐line displays the weighted average in each analysis (i.e., the effect size from column 6 in Table 10).
There are indications of asymmetries primarily for short‐term effects and small‐group instruction. There seem to be more outliers with positive effects and more large than small studies with effect sizes around null. Although they seem less asymmetric, the drastically smaller number of studies in the follow‐up and peer‐assisted categories makes these plots harder to interpret. Conducting Egger's test (Egger et al., 1997; we used the regtest option in metafor), we rejected the null hypothesis of no asymmetry for the short‐term effect sizes (p < .001), and small‐group instruction effect sizes (p = .002) but not for follow‐up effect sizes (p = .165) and peer‐assisted instruction (p = .518). We got qualitatively similar results when we used the “Egger sandwich” test suggested by Rodgers and Pustejovsky (2020): we again rejected the null hypothesis for the short‐term effect sizes (p < .001), and small‐group instruction effect sizes (p = .007) but not for follow‐up effect sizes (p = .107) and peer‐assisted instruction (p = .606).
Asymmetric funnel plots may have other causes than publication bias. In general, asymmetry is a sign of small‐study effects, of which there can be many causes beside publication bias (Sterne et al., 2005). Small‐study effects would show up as heterogeneity (Egger et al., 1997). In line with this idea, the analyses displaying more heterogeneity across effect sizes in Table 10 and that have more studies outside the funnel lines in Figure 14 also have lower p values in Egger's test. A general reason to expect small‐study effects is that researchers often perform statistical power analyses before embarking on an intervention. As interventions with large expected effects require smaller sample sizes for a given level of power, small studies will have larger effects if researchers are reasonably good at guessing the effect sizes (Hedges & Vevea, 2005).
In our context, there are several further reasons why studies with larger sample sizes and consequently smaller standard errors could have smaller effect sizes. One reason is that larger samples tend to be more heterogeneous and may therefore have larger standard deviations (e.g., because they include school district variation and not just school variation as argued by Lipsey et al., 2012). Larger standard deviations would mean smaller standardised mean differences in large studies, even if the effects were exactly the same.
A second reason is that it may be easier to get large positive effects in studies with small samples. Students and teachers can be given more attention and be better monitored by researchers in small studies and there is less risk of coordination and implementation problems. More attention and better monitoring seem likely to increase effect sizes (see Thomas et al., 2018, for an interesting discussion and results in line with this hypothesis) and coordination and implementation problems may cause both smaller effect estimates of otherwise effective programmes and decrease the chances of publication, regardless of publication bias.
A third reason has to do with the recruitment of schools and teachers to studies. In small‐scale studies, researchers typically recruit schools or teachers directly, which implies that only schools or teachers that want to participate are included in the sample. In large‐scale studies, larger administrative units like school districts or municipalities are more likely to be the unit of recruitment. In turn, these larger units may determine which schools and teachers that participate, meaning that some schools and teachers that do not want to participate are included in the sample. If being motivated to participate is important for how well schools and teachers implement the intervention, which seems reasonable (e.g., Kennedy, 2016), then, regardless of publication bias, we should expect smaller effects in large‐scale studies. More generally, it may be easier for small‐scale interventions to select sites that are particularly likely to benefit whereas large‐scale studies is more informative about the population‐wide effects. Depending on the type of policy‐maker (or researcher), both types of effects are interesting, but they are not likely to be the same.
Such connections between sample sizes and effect sizes would violate the assumptions needed to interpret asymmetric funnel plots and Egger's test as publication bias (it would also violate the assumptions underlying the selection models suggested by, e.g., Hedges, 1992 and Andrews & Kasy, 2019). Furthermore, if sample sizes are systematically different across intervention components, the analyses we presented in this review may risk confounding the associations of intervention components with sample sizes.
We examined this issue further by including two study‐level moderators related to some of the reasons stated above in our analyses. The first moderator is equal to one if the effect size/study included a larger than the median number of schools (median = 8), districts (1), or regions (1). If the study contained no information, it was coded as zero along with studies below the median. About 57% of the effect sizes and 52% of studies had a larger number of units regarding at least one of these three units. The second moderator is an indicator for clustered assignment of treatment. As mentioned, 61 studies (around 31%) used a clustered assignment of treatment, which amounted to 17% of the effect sizes (i.e., clustered studies included on average fewer standardised tests/effect sizes). Both of these indicators typically means that the sample size is larger and may imply more coordination problems, as well as a different recruitment process (unfortunately, we did not code how participants were recruited or whether a power analysis was conducted).
We first added the two moderators to the specification reported in column 2, Table 5. As can be seen in Table 9, both indicators were negatively associated with effect sizes but only the indicator for clustered assignment was statistically significant. Reassuringly, none of our other results changed much compared with our primary analysis. We then proceeded to add the same two moderators in the specification underlying Egger's test in the two analyses where this test indicated funnel plot asymmetry. Although the moderators continued to be negatively associated with effect sizes, we still found a significant asymmetry in the funnel plots of overall short‐term effect sizes (p < .001) and small‐group instruction effect sizes (p = .004). As the heterogeneity may depend on outliers, we also ran the test using winsorized effect sizes (with cut‐offs at −0.5 and 1.5). This did not change the outcome of the test, which was still significant for both short‐term effect sizes (p < .001) and small‐group instruction (p = .010).
Of course, these results do not prove that there is publication bias, the moderators are imperfect indicators of the phenomena we want to capture and there may, as discussed, be other reasons for the funnel plot asymmetry.
Funnel plots of study‐level short‐term effect sizes (upper left corner), follow‐up effect sizes (upper right corner), peer‐assisted instruction effect sizes (lower left corner), and small‐group instruction effect sizes (lower right corner)
| Primary analysis | Study‐level effect sizes | |||||||||
|---|---|---|---|---|---|---|---|---|---|---|
| (1) | (2) | (3) | (4) | (5) | (6) | (7) | (8) | (9) | (10) | |
| Analysis | ES | 95% CI | Q | τ2 | I2 | ES | 95% CI | Q | τ2 | I2 |
| Short‐term | 0.3 | [0.25, 0.34] | 797.3 | 0.067 | 76.4 | 0.29 | [0.24, 0.34] | 617.6 | 0.053 | 72.1 |
| Follow‐up | 0.27 | [0.17, 0.36] | 47.8 | 0.031 | 45.6 | 0.25 | [0.16, 0.34] | 35.8 | 0.004 | 8.6 |
| Peer‐assisted | 0.39 | [0.26, 0.52] | 18.7 | 0.023 | 19.8 | 0.38 | [0.24, 0.52] | 11.5 | 0 | 0 |
| Small‐group | 0.38 | [0.30, 0.45] | 285.4 | 0.084 | 70.6 | 0.37 | [0.30, 0.44] | 205.9 | 0.051 | 59.3 |
Summary of sensitivity analysis
The sensitivity analysis showed that the positive and statistically significant overall average short‐term and follow‐up effect sizes, and the positive and statistically significant associations between peer‐assisted instruction and small‐group instruction were generally robust. The exceptions were that there were too few peer‐assisted instruction effect sizes with a relatively low risk of bias on the blinding item for us to include this item in the analysis, and that there were indications of asymmetric funnel plots in the analyses of short‐term and small‐group instruction effect sizes. While we therefore should be more cautious in the interpretation of our results, these results should not be interpreted as a confirmation that there is publication bias. There are other reasons for asymmetric funnel plots, which seem likely to apply in our case.
DISCUSSION
Summary of main results
Our main objective was to assess the effectiveness of targeted interventions for students with or at risk of academic difficulties from kindergarten to Grade 6, as measured by standardised tests in reading and mathematics. We found in total 607 studies that met our inclusion criteria and included 205 of these studies in meta‐analyses: 202 included intervention‐control contrasts and 3 that contained only comparison designs (intervention‐control and comparison design contrasts were never included in the same meta‐analysis). The reasons for not including studies in the meta‐analyses were that they had too high risk of bias (257), that they compared two alternative (and non‐recurring) interventions instead of an intervention and a control group (104 studies), that we were unable to retrieve enough information to calculate an effect size (24 studies), or that the studies used samples that overlapped with other included studies, and had either a higher risk of bias or contained less information (17 studies). Furthermore, 6 studies reported estimates from the same interventions in separate reports/articles, and we treated them as one study cluster in the analysis. Of the 195 study clusters, 327 interventions, and 1334 effect sizes that we included in a meta‐analysis of intervention‐control effect sizes, 93% of interventions were RCTs.
The weighted average short‐term effect size was positive and statistically significant (ES = 0.30, CI = [0.25, 0.34], estimated using 1030 effect sizes and 189 study clusters). The weighted average follow‐up effect size, measured more than 3 months after the end of intervention, was positive and significant (ES = 0.27, CI = [0.17, 0.36], estimated using 195 effect sizes from 27 study clusters). Interventions for students with or at risk of academic difficulties are therefore, on average, effective, and the effects do not disappear immediately after the end of intervention.
All measures of heterogeneity indicated substantial variation among short‐term effect sizes. Follow‐up effect sizes displayed some heterogeneity, but to a much smaller degree. For example, the short‐term τ2 was 0.067 and the I2 was 76.4, and the corresponding statistics were 0.031 and 45.6 for the follow‐up effect sizes. A very large share of follow‐up effect sizes examined small‐group instruction and tested effects on reading measures. We therefore focused the subgroup analysis and investigation of heterogeneity on the short‐term effects.
The subgroup and heterogeneity analyses focused on instructional methods and content domains. We examined average effect sizes in single‐factor subgroup analyses of interventions that included a certain method or domain, and single method and single domain interventions, and recurring combinations of methods, and we used multiple meta‐regressions to examine models that included indicators for methods and domains as well as additional study characteristics.
The main results were that peer‐assisted instruction and small‐group instruction (groups of five students or less instructed by an adult) had large, stable, and statistically significant average effect sizes across specifications (around 0.35–0.45). Both peer‐assisted instruction and small‐group instruction had large effects in studies of interventions where they were the only instructional method used. They furthermore had significantly larger effect sizes than CAI, coaching of personnel, incentives, and progress monitoring in a meta‐regression that included indicators for all coded instructional methods, content domains, and additional study characteristics. Peer‐assisted instruction also had significantly larger effect sizes than medium‐group instruction (groups of six or more students), whereas small‐group instruction did not (p = .07, however). We summarise the results for all components more in detail below, when we compare our results to other reviews. Although all other instructional methods had positive average effect sizes when we analysed interventions including them (and possibly other methods), none was consistently significant across the analyses that tried to isolate the association between a specific method and effect sizes.
The evidence for the average effectiveness of peer‐assisted instruction and small‐group instruction was thus strong, whereas most other methods were examined in only a few single method interventions (CAI and medium‐group instruction were the exceptions with 12 and 10 studies, respectively). However, with the exception of CAI and peer‐assisted instruction in single method interventions, there was still significant heterogeneity within the categories of instructional methods (including substantial heterogeneity in single method small‐group instruction interventions).
We defined peer‐assisted instruction and small‐group instruction broadly. We tried to examine narrower categories, but found only weak evidence of differences. The effect sizes were similar on math and reading tests. Most interventions in the peer‐assisted category examined same‐age peer‐tutoring (or cooperative learning) in pairs. Few examined cross‐age peer‐tutoring or larger groups. The heterogeneity was not significant in single method interventions.
Almost all small‐group instruction interventions targeted academic subjects rather than for example social‐emotional or general academic skills. That is, they were closer to tutoring than mentoring. We found no large or statistically significant differences between instruction one‐to‐one and in groups of two to three, or four to five students. The heterogeneity of the small‐group instruction remained substantial also in these analyses. Using studies that directly compared one‐to‐one instruction with instruction groups of two to five students, we found no significant differences but our statistical power to detect differences was low.
Effect sizes based on math tests was larger than those based on reading tests but the difference was small (around 0.05) and not significant. We found furthermore little evidence that effect sizes were larger in some content domains than others. The exception was interventions targeting fractions, which had significantly higher associations with effect sizes than all other math domains. However, we found only six studies of interventions targeting fractions, and only one of them targeted only fractions. This finding reflected a more general pattern: most interventions targeted more than one content domain, which made it more difficult to isolate the associations between effect sizes and content domains than between effect sizes and instructional methods.
The multiple meta‐regressions revealed few other significant moderators. The mean Grade of the intervention was negatively associated with effect sizes, implying that interventions in higher Grades tend to have somewhat lower effect sizes (a reduction with around 0.03 per Grade). The results for the other study characteristics indicated that effect sizes from QES were not substantially or significantly different from RCTs (we excluded a large share of QES because we assessed them to have too high risk of bias). Furthermore, general tests did not have significantly different effect sizes from tests of subdomains and effect sizes based on math tests were not significantly different from those based on reading tests. In a sensitivity analysis, we used multiple imputation to include variables measuring implementation problems, share of girls, share of minority students, duration, and whether school staff implemented the intervention. None of them was significantly associated with effect sizes.
Overall completeness and applicability of evidence
We conducted a comprehensive search of electronic databases and national indexes/repositories and trial/review archives, combined with grey literature searching, hand searching of key journals, and extensive citation tracking. In addition, we consulted experts in the field. We found a large number of records, which was screened and coded independently by at least two review team members. We searched for studies back to 1980, but included few studies conducted before 2000 and very few conducted before 1990. Therefore, we do not believe that the period limit made us miss a substantial number of relevant studies. We were however unable to retrieve all potentially relevant records in full text (k = 201). For reasons we discuss in the next section, we believe that they are unlikely to have biased our results.
In line with the comprehensive search, we included many forms of publications: journal articles, working papers, conference papers, dissertations, and government reports. However, there may still be recent unpublished studies that we did not find, as educational researchers do not have a tradition of publishing working papers. Publication bias may be another source of missing unpublished studies. We discuss this issue further in the next section. We were also unable to include 24 studies that met our inclusion criteria in the meta‐analysis because they lacked information necessary to calculate an effect size, and we were unable retrieve the missing information from the authors. These studies were more often than included studies not published in scientific journals: around half were either reports or dissertations whereas 82% of the studies in the meta‐analysis were published in a scientific journal. The studies with missing information were older than the average included study: around a third were published before 2000 compared with 14% in the meta‐analysis. The country distribution was similar (21 out of 24 were from the United States) and we have otherwise no reasons to expect them to have a different impact than our included studies.
A large share of the studies included in the meta‐analysis were of interventions conducted in the United States. Several other countries were represented among the included studies—Australia, Canada, Denmark, Germany, Ireland, Israel, Netherlands, New Zealand, Sweden, and the United Kingdom—but there were few studies from each of these countries. In a strict sense, our results therefore mainly apply to the United States school context. However, we believe our main results are transportable to other school contexts outside the United States. There are examples of successful targeted interventions in most countries. Instructional methods like small‐group and peer‐assisted instruction are not particularly dependent on the type of school system in place and would in principle be possible to implement in almost any school.
Quality of the evidence
We excluded many effect sizes from the meta‐analyses because our assessment resulted in a rating of too high risk of bias. That is, our assessment was that they were more likely to mislead than inform the analysis. The most common reasons for this assessment were: inadequate (often no) adjustment for confounding factors, confounding of intervention effects with for example school, teacher, class or cohort effects, and non‐comparable intervention and control groups.
A large majority (84%) of the effect sizes with too high risk of bias were from QES. Reasons for excluding effect sizes from RCTs were for example that the randomisation was compromised and there was inadequate control for confounding, because only one unit was assigned to the intervention or control group, because the studies reported results for a subset of included tests or students, or because of large‐scale attrition. QES were not associated with higher effect sizes than RCTs in the primary analysis, the QES made up a small share of the included interventions (7%), and our results were similar when we only included RCTs in the analysis. We therefore do not believe that the inclusion of QES biased our results.
The effect sizes included in the meta‐analysis had a risk of bias that ranged from low to high, with most effect sizes having a moderate to high risk on at least some items. We conducted a sensitivity analysis in which we adjusted for ratings on the risk of bias items. There were no significant associations between risk of bias ratings and effect sizes and our main results were unaffected by the inclusion of moderators based on the risk of bias items. However, the low number of effect sizes from peer‐assisted instruction interventions rated with relatively low risk on the blinding item made it impossible for us to include this item in the sensitivity analysis.
Despite the lack of associations with effect sizes, it is worth discussing in some more detail what caused the high risk of bias ratings and how the risk of bias potentially can be decreased in future studies. Information about how the random sequence was generated was lacking in most RCTs, and the description of the randomisation procedure was often sparse. As such information is easy to include, this is an area where the reporting of studies can be improved.
Blinding was a problem in all included studies. Complete blinding is difficult to achieve in educational research, but it is for example often possible to use testers that are blind to treatment status. Lack of blinding could both bias the results in favour of the intervention group and in favour of the control group (Glennerster & Takavarasha, 2013). For example, if knowledge about treatment status and that they are participating in an experiment make students try harder—that is, a Hawthorne effect—then the beneficial effects are overstated. However, control group students (or their parents or teachers) may seek out help elsewhere or try harder because they know they did not get the intervention or because they want to compete with the intervention group (i.e., a John Henry effect). In that case, the beneficial effects are understated. We were unable examine this issue with the material at hand, and we are not aware of any other study that have examined this issue in educational interventions.
Around 25% of effect sizes had a low risk of bias (rating 1) in terms of incomplete outcomes. If attrition by comparatively low‐achieving students in the intervention group is more common, then the effects in our meta‐analysis would be overestimated. However, it seems plausible that successful interventions may also make low‐achieving treated students stay in school or show up at testing occasions at a higher rate than the control group. Such a pattern would instead imply that the effects were underestimated. Incomplete outcome data is of course difficult to avoid completely. Nevertheless, some studies could do more to mitigate these problems by examining and testing whether there is differential attrition between intervention and control groups, and adjust for such attrition if present. The data needed to perform such tests and adjustments are usually available to study authors.
Skewed data increase the risk of bias when analysing continuous outcomes, particularly in small sample studies. Consequently, such data may bias our meta‐analysis. We found little evidence of problems with skewed data in our risk of bias assessment, although one reason may be that relatively few studies provided information about more moments of the distribution of outcome variables than means and variances. Another problematic feature of the included studies is the near universal lack of pre‐published protocols and analysis plans. This made it difficult for us to assess whether there was selective reporting or not, but, more importantly, pre‐publishing trial protocols and analysis plans could also mitigate researcher bias and promote transparency.
Besides including the risk of bias items as moderators, we tested the sensitivity of our results to how effect sizes were measured, to outliers, by adjusting for the clustered assignment of treatment, and by including moderators with missing observations. Our main results were robust across these sensitivity analyses.
We also tested if there was heterogeneity across the control group's progression from pre‐ to posttest. One interpretation of such heterogeneity is that the quality of control group instruction differs across interventions, which in turn would be a source of bias. As we were unable to develop moderators based on the control group instruction, such heterogeneity may also explain heterogeneity across effect sizes. While we found strong indications of heterogeneity, there were few significant associations between intervention components and the control group progression.
We found some indications of asymmetric funnel plots for the studies included in the analyses of short‐term and small‐group instruction effect sizes. We performed a thorough search for studies not published in scientific journals and we found only small and not significant differences between effect sizes from studies published in journals and from studies published elsewhere. A possible interpretation of these results is that the missing effect sizes are mainly a file‐drawer problem (Rosenthal, 1979). This interpretation is consistent with the evidence presented in Franco et al. (2014), which indicates that the file‐drawer problem is the main culprit behind publication bias in the social sciences.
There are many other possible explanations for the asymmetric funnel plots, which do not involve publication bias. We believe at least four reasons why small studies tend to show larger effects may be pertinent in our case: First, statistical power analyses may generate a connection between sample size and effect sizes. Second, larger samples tend to be more heterogeneous, and may therefore have larger standard deviations and smaller standardised mean differences. Third, small samples makes monitoring, implementation, and coordination easier. Fourth, larger studies may be more likely to recruit schools and teachers that do not volunteer to participate.
Potential biases in the review process
We performed a comprehensive search and all records were screened and coded by at least two independent screeners in order to minimise the risk of bias in the review process. Three features of the process are however worth discussing in some more detail as they may be a cause for concern.
First, the review team has included many people during the screening and coding phases, which increases the risk of inconsistencies. All team members were thoroughly introduced to the review methods used, and extensive pilot screening and coding were undertaken in each case. All uncertain cases during both first and second level screening were assessed by at least one review author, in addition to two review team assistants. The number of people involved in the coding and assessment of studies was smaller, which should increase the level of consistency. For example, at least one of the first two authors assessed risk of bias for all studies.
Second, we were unable to retrieve 201 records in full text, which amounts to 5% of the total number of records screened in full text, and 0.8% of the total number of records. A minority share of these records likely pertains to another review about students in Grades 7–12, which shared the search and screening process with this review but included less studies (247 compared with 607; see Dietrichson et al., 2020). Furthermore, the records were such that we could not exclude them on being obviously irrelevant in the first level screening, not records that necessarily were relevant. Around 5% of the studies screened in full text were included in any meta‐analysis in this review. Furthermore, older reports from the 1980s and dissertations were overrepresented among the potentially missing studies, which were types of studies that less often met our inclusion criteria. Due to these features, we believe that very few of the 201 not retrievable studies would have been included in our analyses, and that the risk that our results are biased because of these missing records is low.
Third, most of our included interventions were implemented in English‐speaking countries, and 86% were from the United States. Although our search was not limited to records written in English and we did find studies in other languages, we had to restrict the included studies to languages that the review team understood (Danish, English, German, Norwegian, and Swedish). As a result, we may have missed studies from countries where none of these languages are used. Another reason for the dominance of studies from English‐speaking countries could be the, at least historically, stronger focus on qualitative methods in educational research in some European countries (e.g., Pontoppidan et al., 2018).
Agreements and disagreements with other studies or reviews
All reviews including students with or at risk of academic difficulties in kindergarten to Grade 6 that we are aware of have found positive average effect sizes. Furthermore, most reviews that compared different intervention types found substantial heterogeneity of effect sizes. In that sense, our overall short‐term results are in agreement with other reviews.
We are not aware of another review that have provided meta‐analytic estimates of medium‐ or long‐term effects of interventions targeting students with or at risk of academic difficulties. Suggate (2016) reviewed the long‐run effects of phonemic awareness, phonics, fluency, and reading comprehension interventions from preschool up to Grade 7 and included both at‐risk and not‐at‐risk students. Suggate reported positive effect sizes in general, but they were on average reduced by 40% at follow‐up compared with posttest. The mean duration between posttest and follow‐up in Suggate's review was around 11 months. This result is relatively close to our result that effect sizes measured between 12 and 24 months after the end of intervention was 0.17, whereas the short‐term effect sizes in studies that provided a follow‐up measurement was 0.40 (i.e., the follow‐up effects were reduced by 58%).
Below, we comment by intervention component on the most closely related reviews that have used meta‐analysis to examine effect sizes based on standardised or non‐researcher‐developed tests in reading and mathematics for similar at‐risk groups and intervention types in a similar group of countries. Some of the definitions of intervention types used in these reviews were however not comparable to ours, and we only comment on those parts that we deemed were comparable. There are also differences across the reviews in how outcomes were measured and how effect sizes were calculated. We provide the most comparable average effect sizes from the reviews below, but the reader should be aware that they may still not be fully comparable with our effect sizes. Furthermore, with the exception of Gersten et al. (2009), none of the reviews mentioned below used multiple meta‐regressions to examine the association of individual intervention components with effect sizes while adjusting for instructional methods, content domains, and moderators based on study characteristics. Most reviews did not explicitly examine single instructional method or single content domains and their results are therefore closest to the results we reported in Table 3.
CAI
We found significant average effect sizes of CAI across interventions that included this component (ES = 0.15) and in single method interventions (ES = 0.13), but smaller and not significant associations in meta‐regressions where we adjusted for other components and study characteristics. Dietrichson et al. (2017) found a combined average effect size in reading and mathematics of 0.11 for CAI interventions targeting students with low SES. Slavin et al. (2011) and Inns et al. (2019) reviewed interventions for struggling readers and similarly defined instructional technology interventions or programmes. They found average effect sizes of 0.09 and 0.05, respectively. Slavin and Lake (2008) found an overall effect size of 0.19 for CAI interventions targeting mathematics and general student populations, but stated that effects were similar for disadvantaged students and students with a non‐majority ethnic background. Our results for CAI are thus reasonably close to those from comparable analyses in earlier reviews.
Coaching of personnel
The average effect size of coaching of personnel was around 0.20 and statistically significant in interventions that included this instructional method. We found no evidence that coaching was associated with substantial or significant effects in meta‐regressions or in single method interventions. Combining math and reading outcomes, Dietrichson et al. (2017) found an average effect size of 0.16 for coaching interventions targeting students with low SES. They did not examine single method interventions or included coaching in meta‐regressions. Our results for coaching of personnel are thus close to the comparable analysis in their review (Kraft et al., 2018, also reviewed coaching interventions but did not report results for at‐risk groups).
Incentives
We found a relatively large average effect size in interventions including an incentive component (ES = 0.33). The average effect size was smaller and not significant in single method interventions (ES = 0.05) and in the meta‐regressions. Dietrichson et al. (2017) found an average effect size of 0.01 for incentive interventions targeting students with low SES. Although their coding and analysis methods were most comparable to the one we used to obtain the effect size of 0.33, our result seemed to be driven by the inclusion of interventions that combined incentives with other methods, in particular small‐group instruction. Most of these studies were not included in Dietrichson et al. (2017). Our single method and meta‐regression results are close to their results.
Medium‐group instruction
Instruction in medium‐sized groups had a relatively large and significant average effect size in interventions including such a component (ES = 0.32), but smaller and not significant in single method interventions (ES = 0.10) and in the meta‐regressions. Dietrichson et al. (2017) found an average effect size of 0.24 for a similarly defined category (called small‐group instruction) in their review of interventions targeting students with low SES. As the methods used to derive their result was closest to our ES = 0.32, the results are reasonably close.
Peer‐assisted instruction
Peer‐assisted instruction had a relatively large effect size averaged over interventions that included this component (ES = 0.44), in single method interventions (ES = 0.39), and retained significance and a similar size in the meta‐regressions. Dietrichson et al. (2017) found an average effect size of 0.22 for cooperative learning interventions targeting students with low SES. Slavin et al. (2011) found a large average effect size of cooperative learning interventions for struggling readers (ES = 0.58). Inns et al. (2019) found an average effect size of 0.29 for “classroom approaches”, where four out of five studies were of cooperative learning/same‐age peer‐tutoring programmes. Gersten et al. (2009) found very large effects for cross‐age tutoring (ES = 1.02) in their review of math interventions for school‐age learning disabled students, but they included only two such studies. The effects for same‐age (within‐class) peer‐assisted instruction was lower, 0.14, and decreased further when they adjusted for other methods and study characteristics in a meta‐regression.
The target group in Gersten et al. likely had more severe academic difficulties than the average student in our review (as well as the target group in Slavin et al., 2011, and Inns et al., 2019). One explanation of the differences across reviews is therefore that peer‐assisted instruction may work less well for students with the greatest difficulties. Another possible explanation may be that Gersten et al. also included effect sizes based on non‐standardised tests, but they adjusted for this in the meta‐regression and non‐standardised tests typically yield larger effects. Gersten et al. also included interventions targeting older students (although most of their included studies included participants in our Grade range). Note that Gersten et al. only included math interventions, but our results were similar for peer‐assisted instruction across reading and math tests.
Progress monitoring
Interventions including progress monitoring had a significant average effect size of 0.17, whereas there were only two progress monitoring interventions where this method was the only one (ES = 0.28). The associations with effect sizes was small and insignificant in our meta‐regressions. Dietrichson et al. (2017) found an average effect size of 0.32 for progress monitoring interventions targeting students with low SES, but they included only four studies of such interventions and this method was in all cases combined with other methods. Gersten et al. (2009) examined math interventions for students with learning disabilities. Their category teacher feedback combined (ES = 0.23) was reasonably close to how we defined progress monitoring. As in our case, when Gersten et al. included other moderators in a meta‐regression the teacher feedback components (Gersten et al. split the component in two in 7their meta‐regression) lost their statistical significance.
Small‐group instruction
Small‐group instruction showed consistently significant and relatively large effect sizes in analyses of interventions including this component (ES = 0.38), in single method interventions (ES = 0.38), and retained both size and significance in meta‐regressions. Dietrichson et al. (2017) found an average effect size of 0.36 for tutoring interventions targeting students with low SES in an analysis similar to our Table 3. Inns et al. (2019) found an average effect size for “one‐to‐small group” tutoring of 0.20 and 0.25 for one‐to‐one tutoring. Slavin et al. (2011) found an average effect size for one‐to‐one tutoring by teachers of 0.39, by paraprofessionals of 0.38, and volunteers of 0.16, and an average effect size for small‐group tutoring of 0.31 (both Inns et al., 2019, and Slavin et al., 2011, reviewed programmes/interventions for struggling readers in Grades K‐5/6). Wanzek et al. (2016) found larger average effect sizes of one‐to‐one (ES = 0.50), one‐to‐two or three (ES = 0.61), and one‐to‐four or five (ES = 0.44) student small‐group instruction, but they included only interventions in Grades K‐3 in their review of students with or at risk of reading difficulties. Wanzek et al. (2018) found average effect sizes of 0.59 for one‐to‐one tutoring and 0.33 for small‐group tutoring in the review of intensive (100 sessions or more) reading interventions for at‐risk students in Grade K‐3. The results of our review was thus in between the effect sizes reported in earlier reviews, which may be explained by our broader inclusion criteria and larger number of studies.
Other reviews with a target group consisting of more general student populations have also analysed small‐group instruction and tutoring. As tutoring nearly always is targeting students with or at risk of difficulties, we discuss them as well. Fryer (2017) reported an average effect size of 0.31 for math achievement and 0.23 for reading for “high‐dosage tutoring”, which was quite close to our definition. Pellegrini et al. (2018) found an average effect size of 0.25 for combined one‐to‐one and small‐group tutoring in math. Nickow, Oreopoulos, and Quan (2020) examined RCTs in pre‐K to Grade 12 of tutoring, defined as one‐on‐one or small‐group instructional programming by teachers, paraprofessionals, volunteers, or parents, and found an overall pooled effect size of 0.37. Ritter et al. (2006) reviewed the effectiveness of volunteer tutoring programmes for improving the academic skills of student enroled in Grades K‐8 (in the United States), and found effect sizes of 0.30 for reading outcomes and 0.27 in mathematics. As our small‐group instruction category contained almost only tutoring interventions, these results agreed reasonably well with ours.
Content domains
We found few differences between math and reading effect sizes, and between effect sizes from interventions targeting more narrowly defined content domains. Interventions targeting the reading domains comprehension, decoding, fluency, spelling and writing, and vocabulary had average effect sizes ranging from 0.20 to 0.32. We examined the following mathematics domains: algebra/pre‐algebra, fractions, geometry, number sense, operations, and problem solving, which had average effect sizes ranging from 0.15 to 0.50. Only comprehension (ES = 0.21), decoding (ES = 0.31), number sense (ES = 0.51), and operations (ES = 0.17) had been examined in more than one single domain intervention. The meta‐regressions revealed few significant differences across content domains. The exception was fractions, which had a significantly stronger association with effect sizes compared with the other math domains. However, given the small number of interventions targeting fractions and that just one of them targeted only fractions, we caution against strong conclusions based on this result.
Few reviews have examined effect sizes by the domains targeted by interventions for our target group. Gersten et al. (2009) examined the following math domains: operations, word problems, fractions, algebra, and general math proficiency. The categories with similar names as ours also had similar definitions. Their word problems category was similar to our problem‐solving category and their general math proficiency similar to our multiple math category. They found larger effect sizes for word problem interventions in a meta‐regression, but the differences to other domains were not significantly different.
Regarding reading domains, Wanzek et al. (2016) examined effects of interventions on standardised measures of language and comprehension, in Grade K‐3. They coded studies by the outcomes, not the by the targeted domain. However, it seems likely that interventions target the areas that are tested, so their coding may be reasonably close to ours. They found an average effect size of 0.38 for language and comprehension outcomes among interventions in Grade K‐3. While this was slightly larger than our effect size for comprehension, students in their review were in lower Grades and we found larger effects for the lower Grades.
Scammaca et al. (2015) included no study using standardised tests in 4–5th Grade, which was their analytical category closest to our Grade range. The overall average effects in Scammaca et al. (2015) measured by standardised tests were 0.25 in Grade 6–8, which was close to our overall effect size on reading tests. Their effect sizes reported by reading domain were larger than ours for reading comprehension (ES =0 .47) and word study (ES = 0.68), and reasonably similar for fluency (ES = 0.17) and multiple components (ES = 0.14). These analyses were not grouped by Grade however, and thus included secondary students as well. Flynn et al. (2012) found larger effects on standardised tests for comprehension (ES = 0.73), decoding (ES = 0.43), and word identification (ES = 0.41) than we did, but found negative effects on fluency (ES = −0.29; they included students in Grade 5–6 as well). Both Scammaca et al. (2015) and, especially, Flynn et al. (2012) included substantially fewer studies in the comparable Grades and it is also unclear what instructional methods were included in their analyses.
Interventions targeting more domain‐general skills than reading and mathematics had positive effect sizes in our review. The average effect sizes for general academic skills (ES = 0.21) and meta‐cognitive skills (ES = 0.24) were statistically significant, while social‐emotional skills (ES = 0.24) was not significant. These domains were rarely the only targeted domain and we found no evidence that interventions targeting domain‐general skills had larger effect sizes than interventions that only targeted reading and mathematics domains. However, we included only standardised tests in reading and mathematics in the analysis, and targeting domain‐general skills may have important effects on other types of tests.
Few of the earlier reviews with similar target groups as ours included similar categories of domain‐general skills. Dietrichson et al. (2017) included a category called psychological/behavioural interventions (ES = 0.05), which was somewhat similar to a combination of our meta‐cognitive and social‐emotional categories. Pellegrini et al. (2018) included a social‐emotional learning category but their category included whole‐school reforms, which were not included in our review, and their review included interventions targeting general student populations. Thus, their average effect size of 0.03 is difficult to compare with ours.
AUTHORS' CONCLUSIONS
Implications for practice
Our results indicate that interventions targeting students with or at risk of academic difficulties from kindergarten to Grade 6 have on average positive and statistically significant short‐term and follow‐up effects on standardised tests in reading and mathematics. We believe these average effect sizes are of an educationally meaningful magnitude. Both short‐term and follow‐up effects were larger than the 0.25 standard deviations deemed “substantively important” by What Works Clearinghouse (2014). They are in between the 70th and 80th percentile of the distribution of effect sizes from RCTs of educational interventions evaluated on standardised tests presented in Kraft (2020). Both effect sizes correspond to around a 58% chance that a randomly selected score of a student who received the intervention is greater than the score of a randomly selected student who did not. The short‐term average effect size was around 50% of the estimated average progression of the control groups in the studies in our sample that provided this information. That is, compared with this estimate, the intervention groups progressed on average 50% more than the control groups during the intervention period.
The average effect sizes are around 30%–50% of the gaps in fourth grade between low and high SES students, and between majority and minority students, in the United States (the proportion depends on subject, see Hill et al., 2008, and Lipsey et al., 2012). Although this comparison should not be interpreted as targeted interventions necessarily reducing the achievement gaps by these proportions (see e.g., Kraft, 2020, for a discussion), we believe the magnitudes imply that targeted school‐based interventions of average effectiveness tend to have meaningful impacts on the gaps. At least in the short‐term and if given only to at‐risk students, the most effective interventions in our sample have the potential to eradicate the gaps.
Educational policy makers either have to choose a specific programme, or design their own (e.g., because evidence‐based programmes are not available in their country). While our review did not use programmes as the basis for the analysis (see Inns et al., 2019 and Pellegrini et al., 2018, for such reviews), our results indicate that peer‐assisted instruction and small‐group instruction were significantly associated with effect sizes in all analyses, and in both mathematics and reading. They were associated with significantly larger effect sizes than almost all other instructional methods in meta‐regressions. Peer‐assisted and small‐group instruction are thus likely to be effective components of reading and mathematics programmes targeting students with or at risk of difficulties. Moreover, small‐group instruction had relatively large effect sizes also at follow‐up measurements, meaning that the effects are unlikely to fadeout immediately after the end of intervention.
We did not find robust evidence that other intervention components (the instructional methods or the targeted content domains) were, on average, associated with positive effect sizes when used on their own. Most components were significant when we analysed interventions that included them as one part. However, as these interventions also could include other components, this analysis cannot identify the effects of the separate components.
Three things are important to note about the interpretation of these results. First, we found few single method and domain interventions. For example, among the instructional methods, only a handful of studies examined coaching of personnel, incentives, and progress monitoring in single method interventions. Fractions did have significantly stronger associations with effect sizes than all other math domains. However, as only one intervention targeted only fractions, more research is needed.
Second, with the exception of fractions, we found no evidence that any of the reading, math, or general content domains were significantly associated with effect sizes when we adjusted for other components and study characteristics. This does not mean that interventions did not improve the achievement in the targeted domain, just that content domains were not important predictors of the effect size. A possible interpretation is that the achievement of students with or at risk of academic difficulties can be improved in all domains targeted by interventions. However, our analysis included a moderator based on the targeted domain in the interventions, not the tested domain. It may be the case that interventions targeting, say, reading comprehension failed to improve comprehension but instead improved other domains that were not explicitly targeted in the intervention but were included among the outcomes.
Third, with the exception of single method CAI and peer‐assisted instruction, our analyses showed both substantial and statistically significant heterogeneity of effect sizes. Our results are (weighted) averages and there may thus be highly effective interventions using a particular component, which are “hidden” as the average also contains ineffective interventions. While the average small‐group instruction intervention was effective, some interventions in this category did not have any effects for reasons that we could not fully explain.
We found larger effect sizes in the lower Grades. Although this result is in line with calls for early interventions (e.g., Heckman, 2006), we want to stress that our results do not provide a strong basis for choosing between earlier and later interventions. As intervention types may differ across the Grades in ways we could not adjust for, the moderator analysis does not provide causal estimates of the effect of implementing interventions in different Grades.
Furthermore, the choice between earlier and later interventions should also take into account the long‐term cost‐effectiveness of interventions. Examining the costs of interventions was outside the scope of this review (but few studies included information about costs). Although we found positive and significant follow‐up effects, evidence about long‐term effects was scarce in our review. Therefore, evidence about long‐run cost‐effectiveness was also scarce. This lack of evidence should not be confused with a lack of cost‐effectiveness. We do not know whether the short‐ and medium‐term effects found in many targeted school‐based interventions are long‐lasting. We also do not know why there is fadeout of effects (Bailey et al., 2017; Kang et al., 2019), and therefore do not know the long‐term cost‐effectiveness of interventions.
A final caveat is that our review does not provide information about which interventions, and intervention components, that are suitable for particular groups of students with or at risk of academic difficulties. As our definition of the target group was broad, it contains students with different severity of difficulties. This is for example pertinent for peer‐assisted instruction. For this method, our results, and the results of other reviews, did not agree with the results in Gersten et al. (2009), who studied math interventions for learning disabled students. Learning disabled students was but one of the groups we included and this group has likely more severe difficulties than the average student in our target group. Gersten et al. found small and statistically insignificant average effect sizes of same‐age peer‐tutoring, the dominant type of peer‐assisted instruction in our review. As hypothesised by, for example, McMaster, Fuchs, Fuchs, and Compton (2005) and Slavin et al. (2011), some students may need more intensive support than peer‐assisted instruction. McMaster et al. (2005) was the only study testing this hypothesis in our sample and we could not analyse the question further.
The question of whether some interventions are more or less effective for certain target groups has a parallel to our discussion of group sizes in small‐group instruction. We found highly similar effect sizes in one‐to‐one and small‐group interventions. While we were able to meta‐analyse interventions in which only the group size differed between the intervention groups, the number of studies was so few that the analysis did not yield reliable results. We therefore do not know whether reducing group sizes increase effect sizes and whether students with more difficulties benefit more from smaller group sizes.
Implications for research
While the research literature on the effects of targeted school‐based interventions to students with or at risk of academic difficulties have grown a lot in the last two decades, there is still much to learn about how to design and implement effective interventions. We found substantial heterogeneity in most of our analyses. For example, effect sizes varied considerably within categories of single method/domain interventions, and there was systematic variation also after adjusting for a large set of intervention and study characteristics in meta‐regressions. Our discussion about group sizes in the previous section highlights that more research is needed about relatively basic features of otherwise well‐studied instructional methods such as small‐group instruction. More studies comparing single features of programmes, such as the group size, while keeping all other programme components constant between the treatment and control group would be important additions to the literature. Similarly, more research comparing the effects of the same programme for well‐defined target groups, for example in terms of severity of difficulties or age, would be interesting.
The design of interventions would also benefit from more knowledge of the details of the interventions than we were able to include in our analysis. The number of recurring combinations of instructional methods was for example few. Future studies could draw inspiration from our results and how they relate to pedagogical and psychological theories of learning and instruction. For example, we found a large average effect size in the few interventions that combined small‐group instruction with incentives. It is possible that interventions have larger effects, if they combine for example rewards that reinforce desirable behaviour, as emphasised by social learning theory, with a method that includes other potentially effective features, as emphasised by cognitive developmental theory and pedagogical theory. Peer‐assisted and small‐group instruction interventions have many features that seem beneficial from the perspectives of social learning theory, cognitive developmental theory, and pedagogical theory. They include rapid feedback and instruction that can be tailored to individual students, and they train regulation of behaviour by for example interaction with role models. Such interaction may also improve the development of higher‐order cognitive functions such as learning how to learn. Most other instructional methods we examined do not tick more than one or two of these boxes. In this sense, our results were well aligned with the three theories.
However, CAI can include rapid feedback and tailor‐made instruction, and can include regulation of behaviour through for example rewards. There should furthermore be no doubt that children can be highly motivated to play computer games. Despite these potential strengths, single method CAI interventions had a relatively small (but significant) average effect size and a significantly smaller association with effect sizes than both peer‐assisted and small‐group instruction in meta‐regressions. The most salient difference between CAI and both peer‐assisted and small‐group instruction is the amount of social interaction. In terms of the theoretical perspectives, being instructed by adults or peers, and giving instruction to peers (in cooperative learning interventions) may facilitate higher‐order learning better than computer‐assisted instruction. Being seen and encouraged by adults or peers may improve self‐efficacy and motivation in a way that may be difficult to achieve without social interaction. It would therefore be interesting to see studies of interventions that combine CAI with a social interaction component.
We had difficulties coding some aspects of the interventions. Most pertinent, we believe, are the severity of the difficulties facing the target group and the quality of the control group instruction. As for example, tests differ, comparing the severity of difficulties across contexts is problematic. Comparing treatment and control group test scores to norms that are representative for the student population in the country would be one way to make the severity assessment easier. Another way would be to include more information about risk group status. It was often difficult to classify the control condition, as it was described in much less detail (or not at all) than the intervention condition in many of the included studies. It may often be much more difficult to get precise information about the control group condition, but such information is essential for the interpretation of effect sizes. We believe these two aspects are important sources of the unexplained heterogeneity found throughout the analyses.
The risk of bias of effect sizes included in this review was, in general, high. We excluded a large number of effect sizes from the meta‐analyses, because of our assessment that they had too high risk of bias. Although some of this risk is difficult or costly to fully mitigate in educational research, we believe it is important to improve research designs. Many QES did not show balance tests and did not adjust for any confounders. Another common reason for too high risk of bias ratings was that studies assigned only one unit (e.g., a school, teacher, or class) to the intervention group or the control group, in which case the intervention effect is likely to be confounded with “unit”‐effects. In both these cases, research designs can be improved using relatively small means. There are also several other steps that researchers can take to decrease the risk of bias. Examples include more detailed reporting about how the randomisation, or more generally the assignment of treatment, was done (including assessments of the risk of selection into treatment); using external testers that are blind to treatment status; testing for differential attrition between intervention and control groups; and pre‐publishing protocols and analysis plans.
We believe two more aspects of study characteristics are worth mentioning. First, although the number of studies from other countries has increased during the last years, the literature is still dominated by studies from the United States. Second, as mentioned, studies following students over extended periods after the end of intervention are still rare. The growing literature on the long‐term effects of targeted preschool interventions indicates that it is possible to follow participants for much longer periods (e.g., Conti et al., 2016; Heckman et al., 2010; Reynolds & Temple, 2008; Reynolds et al., 2010; Rossin‐Slater & Wüst, 2019). Our results should be encouraging for such studies in the sense that the results indicate that there may be longer lasting effects. Long‐term effect estimates would also make it possible to examine the ratio of benefits‐to‐costs and thereby give educational policy makers a better basis for choosing interventions to improve the achievement of students with or at risk of academic difficulties.
ROLES AND RESPONSIBILITIES
Dietrichson, Bøg, Eiberg, Filges, and Anne‐Marie Klint Jørgensen contributed to the writing and revising of the protocol. The search strategy was developed by Anne‐Marie Klint Jørgensen. All authors contributed to the writing of the review. The following review team assistants provided valuable help with screening and coding: Anja Bondebjerg, Christiane Præstgaard Christensen, Anton Dam, Ole Gregersen, Astrid Broni Heinemeier, Freja Jørgensen, Caroline Fromberg Kiehn, Ida Lykke Kristiansen, Erika Lundqvist, Julie Schou Nicolajsen, Vivian Poulsen, Ida Scheel Rasmussen, Tróndur Møller Sandoy, Ida Skytt, Mette Trane, Mai Tødsø Jensen, and Amanda Weber. Jens Dietrichson will be responsible for updating this review as additional evidence accumulates and as funding becomes available.
CONTRIBUTIONS OF AUTHORS
SOURCES OF SUPPORT
Internal sources
VIVE—The Danish Center for Social Science Research.
External sources
No sources of external support.
DECLARATIONS OF INTEREST
Three of the authors were involved in a previous review on a related topic: the effects of interventions targeting low SES students (Dietrichson et al., 2017). Dietrichson and Bøg have co‐authored two studies included in the review, both of them started after the protocol of this review was approved. Dietrichson and Bøg were not involved in the screening, coding, and risk of bias assessment of these two studies. The authors have no vested interest in the outcomes of this review, nor any incentive to represent findings in a biased manner.
PLANS FOR UPDATING THE REVIEW
Jens Dietrichson will be responsible for updating this review, as new studies and additional funding becomes available.
CHARACTERISTICS OF STUDIES
Characteristics of included studies
See Supporting Information Appendicesand. H I
Characteristics of excluded studies
Due to the large number of studies screened in full text, we were unable to describe all excluded studies. For a full list of studies excluded in first and second level screening, please contact Jens Dietrichson (). jsd@vive.dk
Characteristics of studies awaiting classification
No studies are awaiting classification.
Characteristics of ongoing studies
We found no ongoing relevant studies.
Supporting information
ACKNOWLEDGEMENTS
The review authors would like to thank the editors Carlton Fong, Sarah Miller, and Sandra Wilson, the Campbell methods peer referee, Emily Tanner‐Smith, and the anonymous external content and methods peer referees for valuable and insightful comments on the methods and content of this review. We would also like to thank Douglas Fuchs, Lynn Fuchs, Russell Gersten, Nancy Scammaca, Robert Slavin, and Sharon Vaughn, who answered our request about unpublished studies. We are grateful to John Begeny, Timothy Cleary, Shaun Dougherty, Sandra Dunsmuir, Matthew Dunleavy, Matt Dynarski, Anna‐Mari Fall, June Watters Gothberg, Jonathan Guryan, Stefan Gustafson, Rafael Lara‐Alecio, Mia Finneman Schultz, Susanne Prediger, Fuhui Tong, and Katrina Woodworth for their kind response and help in providing us with information about unpublished studies or additional study data. Thanks to the Heads of SFI/VIVE Campbell, Mette Deding, Hans Hummelgaard, and Lisbeth Pedersen, for continued support and efforts to realise this review. We are grateful to Anne‐Marie Klint Jørgensen who helped develop the search strategy. Mathilde Almlund, Anja Bondebjerg, Christiane Præstgaard Christensen, Anton Dam, Ole Gregersen, Astrid Broni Heinemeier, Freja Jørgensen, Caroline Fromberg Kiehn, Ida Lykke Kristiansen, Erika Lundqvist, Julie Schou Nicolajsen, Vivian Poulsen, Ida Scheel Rasmussen, Tróndur Møller Sandoy, Ida Skytt, Mette Trane, Mai Tødsø Jensen, and Amanda Weber provided excellent assistance at key stages in the production of this review.
Dietrichson, J. , Filges, T. , Seerup, J. K. , Klokker, R. H. , Viinholt, B. C. A. , Bøg, M. , & Eiberg, M . Targeted school‐based interventions for improving reading and mathematics for students with or at risk of academic difficulties in Grades K‐6: A systematic review. Campbell Systematic Reviews. 2021;17:e1152. 10.1002/cl2.1152
Footnotes
REFERENCES
INCLUDED STUDIES
INCLUDED IN THE META‐ANALYSIS