We thank Aiken et al . 1 and Davey et al . 2 for carrying out a re-analysis of Miguel and Kremer (2004). 3 We interpret the evidence from the reanalysis as strongly supporting the findings of positive deworming treatment externalities and school participation impacts.

Aiken et al . 1 usefully correct some errors in Miguel and Kremer. 3,Figure 1 presents the original and updated estimates of the key externality and school participation effects side by side (from their Tables 1 and 5, Appendix Tables VII and IX, and our Supplementary Data and Supplementary Data , available as Supplementary data at IJE online), and shows that results are extremely similar: in addition to direct impacts on worm infections in the treatment vs control schools ( Figure 1 , Panel A), Aiken et al. find externality effects on untreated pupils within treatment schools (Panel B), and externality effects across schools up to 3 km away (Panel C). These effects are significant at P  < 0.05. Similarly, they find direct effects on school participation (Panel E) and within-school externality effects (Panel F) at P  < 0.05, and externality effects up to 3 km away (Panel G) at P < 0.10. The direct effects on both infections and school participation are somewhat larger and more precisely estimated with the updated data (Panels A and E). The updated evidence on within-school externalities implies that a key conclusion in Miguel and Kremer 3 —that individually randomized studies underestimate true deworming impacts—remains valid.

Figure 1.

Deworming treatment effect estimates from the original Miguel and Kremer 3 article (black circles) and updated estimates from the Aiken et al.1 re-analysis (light grey circles), with bars denoting associated 95% confidence intervals. Moderate–heavy intestinal worm infection is the dependent variable in Panels A–D, and the school participation rate is the dependent variable in Panels E–H. The estimated effect is: the difference between treatment schools and control schools in Panels A and E; the within-school externality effect for untreated pupils in the treatment schools in Panels B and F; cross-school average externality effect for schools within 3 km of treatment schools in Panels C and G; and 3–6 km of treatment schools in Panels D and H.

As Aiken et al. point out, most of the errors (rounding, inaccurately labelled statistical significance, or data set updates) are minor. The replication also corrects errors in the original code used to estimate externalities. Miguel and Kremer 3 measured externalities among schools located within 3–6 km that were among the 12 closest schools, rather than all schools, as reported. The externality effect on infections at 3–6 km was significant in the original analysis; the updated estimate is negative but not significant (Panel D). The point estimate on the 3–6 km externality effect on school participation is negative and not significant in both the original and updated analyses (Panel H). The lack of infection externalities at 3–6 km (with updated data) means there is little reason to expect schooling externalities at this distance. Importantly, standard errors on school participation externality estimates at 3–6 km become very large with the updated data, nearly doubling.

We disagree with Aiken et al .’s claim in their Discussion that there is ‘no evidence of a total effect’ of deworming on school participation. This assertion is based on an approach that adds unnecessary ‘noise’ to the estimation. An estimator for overall externalities that goes out beyond 3 km, and that puts extensive weight—due to the large numbers of schools at that distance—on the not significant 3–6 km externality estimate makes the overall estimate far less precise. In Appendix A (available as Supplementary data at IJE online) we show that, under reasonable assumptions, the estimator that excludes the 3–6 km externalities is preferred under the standard criterion of minimizing mean squared error (also see Aiken et al . 4 and Hicks et al.5 ).

Since the estimated direct effect ( Figure 1 , Panel E) is a lower bound on overall deworming impacts as long as treatment spillovers are either zero or go in the same direction as the direct effect, the appropriate conclusion is that deworming has a large positive impact on school participation. Combining the precise direct effect estimate and 0–3 km externality estimate with the effect at 3–6 km might have made sense with the original data, but with the updated data the 3–6 km effect estimate is simply too noisy to be informative for statistical inference. Aiken et al . also claim that there is little evidence for treatment spillovers across schools. Yet the finding that externality effects at 3–6 km are imprecisely estimated in no way negates the positive effects within schools or across schools within 3 km (Panels C and G).

The argument in Davey et al ., 2 that deworming impacts on school participation are not robust to different statistical approaches, is based on several analytical errors.

First, they misclassify pre-treatment control observations as treatment observations. Group 2 schools began receiving deworming in March 1999. The correct coding of treatment for Group 2 thus begins after March 1999, as in Miguel and Kremer 3 and Aiken et al .; 1 however, Davey et al . 2 misclassify the Group 2 observations from early 1999 as treatment observations. They purport to justify the misclassification of 20% of 1999 observations using an ‘intention-to-treat’ framework, a framework typically utilized when a population was assigned to treatment, but only some individuals actually received it. Davey et al. incorrectly apply it to a different situation, in which no individuals were actually treated (i.e. Group 2 prior to March 1999) nor were any supposed to be treated. Their puzzling approach rests on the assertion that the programme sought to provide deworming at the exact start of the calendar year. However, the study’s research design necessitated treatment not starting immediately at the start of 1999: extensive data collection was carried out in early 1999 precisely because Group 2 had not yet been phased into treatment, allowing for estimation of impacts after 1 year of treatment (further discussion in Appendix B, available as Supplementary data at IJE online; also see Davey et al.6 and Hicks et al . 7 ).

Second, since all estimators in Davey et al.2 are based on treatment vs control school differences in a context with positive treatment externalities, their estimates are biased towards zero. Furthermore, many of the estimators ignore the study’s stepped-wedge design, in which some schools change treatment status in year 2. They instead focus on cross-sectional estimates, and moreover, split the data into year subsets and report results separately for the subsets, unnecessarily sacrificing statistical precision. The re-analysis authors’ own power calculations imply that this approach is extremely underpowered (Aiken et al. , 8 Davey et al. , 6 Appendix 1, available as Supplementary data at IJE online). An analysis using both years of data and the stepped-wedge design—the specification that represents the culmination of their own analysis plan (Aiken et al . 8 )—produces large impacts; in their abstract they write: ‘When both years were combined, there was strong evidence of an effect on attendance’.

Figure 2 shows that the attendance result is robust across a range of approaches common in economics and public health. These estimates employ both years of data, as envisioned in the project’s prospective design, and use: (i) different statistical models (linear regression, random effects logistic regression); (ii) different samples (the full sample, those eligible for treatment); (iii) regression models adjusted for covariates and unadjusted; (iv) different approaches to weighting (each attendance observation equally, each pupil equally); and (v) the dataset in Davey et al. that incorrectly defines treatment, vs data that correctly defines treatment. All 32 estimates in Figure 2 are positive, large and significant ( P  < 0.01) (details in Supplementary Data , available as Supplementary data at IJE online). The pre-specified logistic analysis in Aiken et al. ’s 8 plan is represented by a bold vertical line in Panel B, and this estimate of 1.82 is significant at P  < 0.001.

Figure 2.

Deworming treatment effect estimates on school participation. Each vertical grey line denotes a coefficient estimate of the effect of deworming on school participation. The estimates use both years of data, and differ in: (i) statistical model (the original linear regression model in Panel A, and random effects logistics regression from Davey et al . 2 in Panel B); (ii) sample (the original full sample, and the sample eligible for treatment in Davey et al .); (iii) regression models adjusted for covariates and unadjusted; (iv) approaches to weighting observations (each attendance observation equally, and each pupil equally); and (v) the dataset that in Davey et al . employ in their analysis, which incorrectly defines treatment and makes additional missing data assumptions (Appendix B), vs data that correctly define treatment. All 16 coefficient estimates in Panel A are significant at P  < 0.01; all 16 estimates in Panel B are significant at P  < 0.001. The bold vertical lines denote the adjusted model estimate using Davey et al. ’s 2 data; the Panel B estimate is from their Table 2, top right panel.

To justify not pooling both years of data, Davey et al . 2 raise concerns about the correlation between the number of attendance observations per school and school participation rates, in the treatment vs control schools over time, which they apparently establish by ‘eyeballing’ a plot of the relationship. We present statistical evidence that this correlation is not significant and does not bias estimates ( Supplementary Data , available as Supplementary data at IJE online). Davey et al . also base part of their conclusion on a cluster-level analysis using a non-standard approach to weighting observations. Even in the cluster-level models, we show that deworming has a significant positive effect on school participation for each year separately when standard weighting is applied and treatment is correctly defined ( Supplementary Data , available as Supplementary data at IJE online). It is only when one simultaneously makes multiple analytical errors—in weighting observations, defining treatment, and failing to pool the data—that deworming impact estimates are not significant.

In sum, a re-analysis of Miguel and Kremer 3 confirms its main conclusions regarding deworming treatment externalities and school participation gains ( Figure 1 ), and these school participation gains are robust across a wide range of adequately powered analysis methods ( Figure 2 ). These results contribute to a growing literature using cluster randomized or quasi-experimental designs to study deworming’s socioeconomic impacts, all of which estimate positive long-run impacts on educational and labour market outcomes (Ahuja et al.,9 Bleakley, 10 and Ozier 11 ).

Acknowledgements

We thank Kevin Audi, Evan DeFilippis, Felipe Gonzalez, Leah Luben and especially Michael Walker for excellent research assistance. All errors remain our own.

Conflict of interest: MK is Scientific Director for Development Innovations Ventures at USAID. Among its many other activities, USAID supports deworming.

References

1

Aiken
A
Davey
C
Hargreaves
J
Hayes
R
.
Re-analysis of health and educational impacts of a school-based deworming program in western Kenya: a pure replication
.
Int J Epidemiol
2015
;
44
: 1572
80
.

2

Davey
C
Aiken
A
Hayes
R
Hargreaves
J
.
Re-analysis of health and educational impacts of a school-based deworming programme in western Kenya: a statistical replication of a cluster quasi-randomized stepped-wedge trial
.,
Int J Epidemiol
2015
;
44
:
1581
92
.

3

Miguel
E
Kremer
M
.
Worms: identifying impacts on education and health in the presence of treatment externalities
.
Econometrica
2004
;
72
:
159
217
.

4

Aiken
A
Davey
C
Hargreaves
J
Hayes
R
.
Re-Analysis of Health and Educational Impacts of a School-Based Deworming Program in Western Kenya: Part 1: Pure Replication
.
3ie Replication Paper 3, part 1. Washington, DC: International Initiative for Impact Evaluation (3ie), 2014
.

5

Hicks JH, Kremer M, Miguel E. (a) Estimating Deworming School Participation Impacts and Externalities in Kenya: A Comment on Aiken et al. Original author response to 3ie Replication Paper 3, Part 1. Washington, DC: International Initiative For Impact Evaluation (3ie), 2014.

6

Davey
C
Aiken
A
Hayes
R
Hargreaves
J
.
Re-Analysis of Health and Educational Impacts of a School-Based Deworming Program in Western Kenya: Part 2: Alternative Analyses
.
3ie Replication Paper 3, Part 2. Washington, DC: International Initiative For Impact Evaluation (3ie), 2014
.

7

Hicks JH, Kremer M, Edward M. (b). Estimating Deworming School Participation Impacts in Kenya: A Comment on Aiken et al . Original author response to 3ie Replication Paper 3, part 2. Washington, DC: International Initiative for Impact Evaluation (3ie), 2014.

8

Aiken
A
Davey
C
Hayes
R
Hargreaves
J
.
Deworming Schoolchildren in Kenya – Replication Plan
.
Washington, DC
:
International Institute Impact Evaluation
(3ie), 2013
.

9

Am
Ahuja
Baird
S
Hamory Hicks
J
Kremer
M
Miguel
E
Powers
S
.
When Should Governments Subsidize Health? The Case of Mass Deworming
.
National Bureau of Economic Research Working Paper #21148. Cambridge, MA: National Bureau of Economic Research, 2015
.

10

Bleakley
H
.
Disease and development: evidence from hookworm eradication in the American South
.
Q J Econ
2007
;
122
:
73
117
.

11

Ozier
O
.
Exploiting Externalities to Estimate the Long-Term Effects of Early Childhood Deworming
.
World Bank Policy Research Working Paper #7052, Washington, DC: World Bank, 2014
.

Supplementary data