Hi Tony, This is Mark again. Thanks for the nudge on this monitoring point!
We agree that real-time monitoring of critical activities is likely to be a cost-effective investment. For example, New Incentives (a program that provides conditional cash transfers for immunizations working with Northern Nigeria health clinics) tracks and shares with us data on cold chain failures, stockouts, rollout delays, security incidents, fraud, double treatment of children, and retention rates (see more information in this intervention report). We think this type of monitoring is a good investment for New Incentives and want others to do something similar, because it:
We're still trying to understand the trade-offs between grantees doing this internally vs. externally, and we expect to learn as we go. It makes sense to me that it is usually most efficient and effective for grantees to do the core of this monitoring internally since we want them to be able to address these issues in real-time (point 1) rather than waiting for a report from an external firm. In cases where GiveWell is relying on the accuracy of this data for our decision-making (points 2 and 3), I think it makes sense to double-check data accuracy through external data quality audits and/or interviews with local stakeholders.
The same goes with standardizing. We think it's good to have a minimum standard, but we guess there's also variation in what's feasible or makes sense across grantees and expect to figure that out with them. We’ve standardized monitoring and reporting within a few program areas, such as across all our recent water chlorination grants, in collaboration with our grantees (e.g., agreeing on consistent standards on what grantees are expected to track, at what frequency, and how findings should be acted on). However, we haven’t developed monitoring standards that apply across all program areas. I think there could be value in having broader principles, but, so far, we’ve prioritized addressing concrete issues we’ve identified rather than developing these broader principles. We are planning to publish more on the details of these program area-specific monitoring and reporting standards in the near-future.
Hi Tony—This is Mark Walsh, a GiveWell researcher on the team responsible for pressure-testing GiveWell's research processes and conclusions. Thank you for this thoughtful write-up!
At a high level, I agree with the central thesis: we've underinvested in monitoring and evaluation, relative to other components of our analysis. Like you mentioned in your post, we've been working to fix that, and I wanted to share a little bit more on what we've done so far, what we're planning, and some other related gaps in our work.
We think this kind of external engagement with our research is valuable and makes our work better. We'd welcome feedback on the steps we've taken so far and where we should consider doing more.
What we’ve done
Over the past year, we've been working on what we call "M&E red teaming"—a systematic review of the monitoring and evaluation practices of our largest grantees, motivated by the same concerns Tony raises. We're planning to publish our findings soon but wanted to give a brief overview.
From June to December 2025, we dedicated teams of 3-4 research staff to work full-time for roughly 3 weeks each on six program areas: our four top charities plus water chlorination and malnutrition treatment. For each program, we evaluated many of the dimensions Tony highlights in his monitoring checklist: the independence of data collectors from program staff, the neutrality of the sampling frame, the objectivity and precision of measurement approach, data quality checks and backchecks, timeliness of data, triangulation with independent sources, and whether the program is taking timely action to address any issues raised by monitoring.
Since completing the red teaming, we've been doing the following to make improvements:
As Tony suggests, we think the right way to evaluate these investments is based on the value of the information we’ll gain and the impact we think it will have on our future grantmaking decisions. That means that we are more likely to fund expensive M&E in large grantmaking areas for GiveWell (or areas with a lot of room for more funding) and where we have more uncertainty so expect the M&E to affect our grantmaking a lot. For example, we’ve funded expensive independent coverage surveys of insecticide-treated nets and vaccinations programs because we direct a lot of funding to these programs. On the other hand, we are trying to take a rigorous but lighter-touch approach in cases where there is less funding at-stake, or where key uncertainties (e.g., whether an organization can establish partnerships or hire effectively) can be resolved more cheaply before investing in expensive M&E.
What’s next
Over the next several months, we're planning to finalize and roll out our coverage survey standards with grantees; analyze the results from the independent surveys, enhanced M&E, and population/costs work we've commissioned; and dig further on more of these areas (e.g., investigating better ways to monitor the impact of our grantees on disease morbidity and mortality).
I expect we'll learn a lot as we go about what's feasible and what information is actually most valuable. I'm sure our approach will change as we learn more about both the cost and potential impact of this information. That being said, I think we are on the right track.
A broader gap
I also want to flag that we think the issues Tony raises are part of a broader gap. It's not just that we need better quantitative monitoring—we also need to invest more in understanding what's actually happening on the ground with the programs we fund.
We've been trying to gather more "local insights" on our work. This involves site visits, qualitative research, conversations with local experts, and other ways of testing our desk-based assumptions against what’s happening on the ground. One example is funding the Busara Center for Behavioral Economics to observe vitamin A supplementation delivery in Nigeria and interview households, front-line staff, and government officials about the program.
We're still figuring out which approaches are most useful but think shifting more of our research effort "beyond the spreadsheet" in the ways Tony is describing is directionally right and something we’re making progress on. As I said at the start, we welcome feedback on our work so far—and on our future progress as it occurs.
Hi there - thanks for your interest in GiveWell! We don't currently offer informational calls, but we'd encourage you to check out our open roles and apply to any that seem like a good fit for your background and interests.
Hi Nick,
Thanks for noting that section of the post could have been clearer! We’ve edited the article to clarify that New Incentives went from serving 70,000 to 1.5 million children per year.
We agree that the extra lives saved (“indirect deaths” in our analysis) is an interesting question. Both the magnitude of the adjustment and the exact mechanisms (i.e., which other causes those deaths are coming from in the GBD bucket) are major sources of uncertainty in our model, and we don’t currently specify what other deaths are being averted through vaccination in our analysis. We may follow up with a post to share more about our work on indirect deaths in the future.
Thanks again for the feedback!
Hi Nick,
Thank you for providing this feedback! My name is Vicky, and I am a Research Associate at GiveWell, on the vaccines team. We really appreciate these kinds of rough sense checks on our work and thought this was a great approach.
Our lookback includes children enrolled across multiple years of programming (roughly covering 2020 to 2026) whereas the enrollment figures in your estimate only include a single year of program operations.
We think this difference–the assumed number of children enrolled with GiveWell funding—is the main reason the upper bound you estimated for the number of deaths averted appears significantly lower than the estimates in our lookback, although we’re still exploring other potential discrepancies between the numbers in your approach and our estimates.[6]
Thanks again for your engagement!
In 2023, New Incentives reported enrolling 1,518,904 children across 9 states. See New Incentives, 2023 Annual Report, p.8-9
We estimated this by taking the total amount of funding (roughly $120 million) divided by the cost per child enrolled (roughly $19 per child enrolled) between 2020 and 2024. This assumes that the cost per child enrolled between 2025 and 2026 will remain similar to the historical weighted average.
The 81% in our public report is based on a single state, Bauchi, and the exact percentage differs across states depending on baseline coverage and New Incentives’ expected impact in that state. In addition, we've made some internal updates to the model since the last version of our intervention report was published.
6.3 million * (1 - 81%) = roughly 1.2 million children counterfactually vaccinated by the program.
1.2 million children counterfactually vaccinated * 5% risk of dying from causes that might be preventable through vaccination = 60,000 deaths potentially averted as an upper bound.
Across states where the New Incentives program operates, we estimate that unvaccinated children experience roughly a 3% to 8% chance of dying from vaccine-preventable diseases and that vaccination reduces their risk of dying by roughly 50%, which appears more in-line with your estimates. For more on how we estimate these, see our public report here.
GiveWell Updates
- GiveWell is launching a new RFI to expand and strengthen our malaria grantmaking in Africa and help our donors make a greater impact. Expressions of interest can be submitted through one of two tracks, the first for malaria chemoprevention and vector control pilot programs and the second for research and evaluation. Submissions are due June 24.
Learn More- Most of GiveWell’s grantmaking focuses on programs that reduce child mortality, but our growing research capacity has expanded what we’re able to evaluate and fund, including highly cost-effective programs that meaningfully improve quality of life. In a recent podcast episode, GiveWell CEO and co-founder Elie Hassenfeld speaks with Program Officer Meika Ball about our grant to MiracleFeet—an organization that expands access to clubfoot treatment—and her recent site visit to see the program in action in Côte d’Ivoire.
Listen Here: Evaluating and Funding a New Kind of Grant- Save the date for our upcoming webinar, Looking Back to Give Better: How GiveWell Evaluates Its Grantmaking, on Tuesday, June 9. GiveWell co-founder and CEO Elie Hassenfeld will moderate a conversation with Program Directors Alex Cohen and Julie Faller about how we evaluate whether a grant achieved the impact we initially estimated—and how we use those findings to make better impact estimates over time.
Register Here- Vitamin A supplementation has one of the strongest evidence bases of any program we've evaluated. But when you dig in, the evidence is more complicated than it looks. GiveWell Senior Researcher Stephan Guyenet joins CEO and co-founder Elie Hassenfeld on the GiveWell Conversations podcast to talk about the evidence for vitamin A supplementation, the hard questions that remain, and how our expanded research capacity is helping us go deeper so we can direct funding more effectively.
Listen Here: Scrutinizing One of Our Longest-Funded Programs