I’m working on a pamphleting pilot program with the folks over at The Life You Can Save, and I wanted to lay out our plans and hopefully field some useful thoughts and criticisms. First, a couple notes on the motivation for this project. Pamphleting has been used by other advocacy groups (with positive, although debatable, results), but to my knowledge hasn't been tried by an EA-flavored global poverty organization, so we're giving it a shot. There is arguably something old-fashioned and appealing about being handed a physical booklet from a real person that differentiates this medium from, say, online ads, and promotes (or so the argument goes) a deeper level of engagement. We're hoping that the data we collect, which TLYCS will be using to decide whether to continue it's program, will be useful to other groups as well, so we wanted to field some feedback here. 

The basic thrust of the project is right out of the Vegan Outreach playbook; put together a leaflet that targets college age students with material relating to global poverty interventions and effective altruism, and hand them out on college campuses on a rolling, semester basis. Whereas for VO the ultimate goal of pamphleting is to make vegans, the goal for the TLYCS pamphlets is to get students to engage with the material on the website and then take TLYCS's pledge. We have our pamphlet nearly finished and plan to start handing them out this spring semester at Southern California universities. In this pilot program, we really want to, 1.) explore the logistical challenges in running this kind of activity, and 2.) try and gauge the effectiveness of pamphleting versus other methods of spreading the message. It is on point two that I am keen to get feedback, so I’ll outline our current thoughts on this.

I plan to hand out pamphlets on discrete days during the pilot period (say, every other friday for several months), and keep track of how many booklets are handed out during each distribution. This should allow us to monitor traffic spikes (or lack of them) to the website from IP addresses in the distribution zone, as well as spikes in pledge signups, which can serve as a preliminary indication of engagement. We will also be pushing a special landing page that is only advertised in the pamphlets as another metric to gauge impact. This would be analogous to VO’s vegetarian starter guide offer, in that the offer require's the reader to actually do something, and therefore ostensibly reflects a more significant level of engagement. We initially envisioned directing students to a “special message” from TLYCS, which might be a video from Peter Singer or someone else in the organization offering an in-person invitation to explore the site and take the pledge. Another thought was to invite students to play an online version of TYLCS’s Giving Game, where they visit the website and choose which of four charities will receive a dollar donation. However, there was some disagreement on the team whether offering a strong “hook” to get people to come to the site would muddy the effectiveness of the pamphlet qua pamphlet. For instance, TLYCS offers in-person Giving Game workshops, and it was thought that pitching an online version of the Giving Game in the pamphlets would make it harder to compare the effectiveness of the pamphlets versus the in-person workshops. As something of a compromise, we are now inviting students to take a quiz (just a shell site right now) in which they answer questions and then get paired up with one of TLYCS’s recommended charities. I’d be very interested in what people think on this point; what’s the right balance between pushing a really strong hook in the pamphlet (e.g. video from Peter Singer, invitation to play Giving Game), versus keeping the pamphlet as independent as possible so we can tell how engaging of a medium it is?

Ultimately, we’d like to come out of the pilot program with a metric that falls somewhere close to “cost per TLYCS pledger via pamphlets” that we can compare with other outreach efforts (online ads, in-person workshops, etc), and use in deciding whether to expand the program or abandon it. To summarize, I’m really interested in hearing comments on:

 

  • Are we making any obvious mistakes in our approach to measuring the effectiveness of the pamphlets?
  • What is the right balance to strike between pitching a strong “hook” in our pamphlets versus keeping them “clean” so that only the appeal of the pamphlets as a medium is measured?
  • Is there anything we can tweak in our pilot that would make it more relevant and useful to other organizations considering a pamphleting program?

Thanks all, and look forward to hearing from you!

 

Comments34


Sorted by Click to highlight new comments since:

What is the right balance to strike between pitching a strong “hook” in our pamphlets versus keeping them “clean” so that only the appeal of the pamphlets as a medium is measured?

What you want to assess is the marginal cost-effectiveness of pamphlets, so I think the right approach is to include the best hook that you can which scales up at zero marginal cost. This probably excludes an online giving game. It should allow an online video, but ideally one that can be reused if distributing these elsewhere rather than too personalised to the target audience.

Really good point here; I was a fan myself of the online Giving Game, but that would be hard to scale with the program without securing a donor willing to finance it at a pretty large level.

If the hook is worth it, how expensive would it be to scale and hard would that be to finance?

I suppose if the initial pamphlet run is worth it, you could then A-B test it with a Giving Games pamphlet.

Hey Jonathon, this is a really great initiative! Giving What We Can is currently in the process of designing an experiment to test the effectiveness of our pamphlets. We were hoping to run it in London some time late January, early February. We should coordinate on our experiment design [I will post more details on the forum once we have firmed up details about the experiment design].

Cool, will the results be public?

Hey Tom, not sure about TLYCS's study, but we plan to make ours public (and I imagine they will too!)

After trailing and fiddling to see what works - how much would 20 million copies of a pamphlette aimed at a general audience cost? The post office gives a lot to charity and I can imagine that it wouldn't be impossible to persuade them to send this out as a one off free of charge - at least to the houses they're already posting mail to.. Perhaps different language for different postcodes? Chelsea does not equal Bradford in terms of how appeals might work (religious backgrounds, education levels, size of household, disposable income etc.)

That would be great! I'll connect with you on Facebook and we can open up a line of communication there.

Also, have you got in touch with the good people at Charity Science?

Just took a look at their website, very cool stuff. You suggesting I email them and get their feedback on our plan?

Definitely. Some of the team at least are EA insiders and lurking on this very forum, and they'll already know about TLYCS for sure.

We lurk amongggggg youuuuu.

Hi Jonathon, will the results be public?

That is definitely the intention. We are really hoping that the data we gather will be useful to other orgs considering a similar program, which was part of the motivation for posting up here ahead of time to get feedback.

There was an "Effective Altruism Brochure" thread on Facebook's Effective Altruists group. Might be a good starting template to use for a handout: see here

Pretty sharp! If I had seen this before, I definitely would have passed it along to our designer as something to work from.

Ultimately, we’d like to come out of the pilot program with a metric that falls somewhere close to “cost per TLYCS pledger via pamphlets”

I noticed what might be a significant confounder in getting this estimate: you are likely to be particularly enthusiastic/eloquent about the whole thing, which is an extra input which will help the effectiveness of the pamphlets, but is very hard to budget properly on the 'costs' side.

To account for this, you would probably be better hiring someone else to distribute the pamphlets; probably someone without a deep existing commitment to TLYCS (but some contact should be fine -- the question is whether you could get similarly good people when trying to scale it up).

But if you're going to scale this, you'll probably get LYCS members or EAs to hand out the pamphlets anyway, right? I mean, we kind-of do need more concrete volunteery tasks that we can give to student groups anyway.

So it's best to get random EAs or LYCS members to do it in the study, right?

I agree that if the people running the study are also distributing the pamphlets then you end up with bias.

I wasn't sure what the scaling model was, but if there are enough plausible volunteers then this sounds right.

The general point is that you want to try and produce a typical case, not a special case.

This is a really good point. Yeah, the scaling model is to have local TLYCS chapters organizing volunteers to do this as a regular, rolling semester activity. I hadn't really considered myself a confounding variable in this sense, because I'm definitely not a master pamphleteer. I'm an engineer by trade, and if this program takes off, I'll eventually just be another volunteer in the LA area that helps hand out leaflets occasionally. We're also thinking about splitting crews on Friday distribution days - so I would have a crew that hits up two universities, and there would be another volunteer crew hitting up two different campuses. Any thoughts on this?

Great idea!

Does the pamphleting have to be done on Fridays, or can it be done on pseudo-random days? (I'm thinking about distinguishing the signal from the pamphlets from e.g. people spending more time on the Internet during weekends. Pseudo-random spikes might require fancier math to pick out though, and of course you need to remember which days you handed out pamphlets!

Can you ask people, when they take the pledge, how they found out about TLYCS? (This will provide an under-estimate, but it can be used to sanity-check other estimates). (Also it's a bit ambiguous if someone had e.g. vaguely heard of TLYCS or Singer before, but pamphleting prompted them to actually take the pledge)

There's a typo in your text ("require's") - make sure you get the pamphlets proof-read :)

Do you know in advance what you expect, in terms of:

  • How many pamphlets you will distribute
  • What the effect will be?

(Last I heard, EA was using predictionbazaar.com and predictionbook.com as its prediction markets)

Statistically, the situation you don't want to get into is leafleting every Friday so there's no Fridays left to provide your control condition.

Oh yeah, good point.

Some really good points here. I never considered that handing out the leaflets only on Fridays might skew the results (I just happen to have every other Friday off, thanks California), I'll have to think that through. And it would definitely be a good idea to have a "Where did you hear about the pledge?" question on the pledge site, I'll check into that as well.

I'm not sure what our initial run on the pamphlets will be, but I'm thinking in the 5K-15K range. I haven't done any analysis to figure out how many we'd need to hand out to get good statistics; not even really sure how to go about doing that, to be honest. And absolutely no idea what to expect in terms of a response rate. Any thoughts on how to estimate that?

I'm not sure what our initial run on the pamphlets will be, but I'm thinking in the 5K-15K range. I haven't done any analysis to figure out how many we'd need to hand out to get good statistics; not even really sure how to go about doing that, to be honest.

Please talk to a real statistician if you're designing an experiment! Random Internet people picking your design apart is actually pretty good as far as review goes (if they're the right Internet people), but actual statisticians are orders of magnitude better. Experiment design is very tricky and good statisticians are aware of both lots of tools to make your job easier, and lots of pitfalls for you to avoid. To quote Ronald Fisher:

To consult the statistician after an experiment is finished is often merely to ask him to conduct a post mortem examination. He can perhaps say what the experiment died of.

Statistics Without Borders may be a good place to start.

I figure most people don't know a statistician (I don't) but there are plenty of people in LessWrong discussion who know how to do a power calculation so it might be good to start there (or just to dig a bit deeper here).

It really won't help address the problem I'm talking about at all, which is unknown design flaws/statistical techniques/study design tools. Once you've figured out that you have a problem like "how should I power my study?", smart people plus Internet is fine; I'm worried about the other 10 issues we haven't noticed yet. That's the kind of thing that statisticians are useful for.

Fortunately, it turns out you can still talk to statisticians even if you don't know them personally. If you're spending money on your study, you could even go so far as to hire a consultant. I also know statisticians and would be happy to refer Jonathon.

Makes sense. Also, if the EA survey is redone, that might be an even more important place to have a statistician.

As someone who did a lot of study design in undergraduate, is currently a "data scientist", and considers myself smart, I can confirm that I still make approximately 10 huge mistakes every time I run a study.

Yeah, if you give me the contact info of a statistician that you recommend that would be great. I don't know if we have the budget for it, but I would definitely reach out.

I'm checking for people who would be interested in doing it pro bono. If that doesn't work, I'm 99% sure you can find some people to fund a couple consultant-hours.

I don't know if we have the budget for it

Not to put too fine a point on it, but if the alternative is TLYCS designing the experiment themselves, this is pretty much like running a charity that spends nothing on overhead. It looks good on paper, but in reality, that last bit of money is a huge effectiveness multiplier.

I'd consider funding this if it's "worth it" and not too much money. I'm sure others would as well.

I'm fairly surprised the EA movement doesn't have official statisticians. The EAA movement has a lot of people claiming to be official statisticians.

Curated and popular this week
 ·  · 10m read
 · 
Regulation cannot be written in blood alone. There’s this fantasy of easy, free support for the AI Safety position coming from what’s commonly called a “warning shot”. The idea is that AI will cause smaller disasters before it causes a really big one, and that when people see this they will realize we’ve been right all along and easily do what we suggest. I can’t count how many times someone (ostensibly from my own side) has said something to me like “we just have to hope for warning shots”. It’s the AI Safety version of “regulation is written in blood”. But that’s not how it works. Here’s what I think about the myth that warning shots will come to save the day: 1) Awful. I will never hope for a disaster. That’s what I’m trying to prevent. Hoping for disasters to make our job easier is callous and it takes us off track to be thinking about the silver lining of failing in our mission. 2) A disaster does not automatically a warning shot make. People have to be prepared with a world model that includes what the significance of the event would be to experience it as a warning shot that kicks them into gear. 3) The way to make warning shots effective if (God forbid) they happen is to work hard at convincing others of the risk and what to do about it based on the evidence we already have— the very thing we should be doing in the absence of warning shots. If these smaller scale disasters happen, they will only serve as warning shots if we put a lot of work into educating the public to understand what they mean before they happen. The default “warning shot” event outcome is confusion, misattribution, or normalizing the tragedy. Let’s imagine what one of these macabrely hoped-for “warning shot” scenarios feels like from the inside. Say one of the commonly proposed warning shot scenario occurs: a misaligned AI causes several thousand deaths. Say the deaths are of ICU patients because the AI in charge of their machines decides that costs and suffering would be minimize
 ·  · 14m read
 · 
This is a transcript of my opening talk at EA Global: London 2025. In my talk, I challenge the misconception that EA is populated by “cold, uncaring, spreadsheet-obsessed robots” and explain how EA principles serve as tools for putting compassion into practice, translating our feelings about the world's problems into effective action. Key points:  * Most people involved in EA are here because of their feelings, not despite them. Many of us are driven by emotions like anger about neglected global health needs, sadness about animal suffering, or fear about AI risks. What distinguishes us as a community isn't that we don't feel; it's that we don't stop at feeling — we act. Two examples: * When USAID cuts threatened critical health programs, GiveWell mobilized $24 million in emergency funding within weeks. * People from the EA ecosystem spotted AI risks years ahead of the mainstream and pioneered funding for the field starting in 2015, helping transform AI safety from a fringe concern into a thriving research field. * We don't make spreadsheets because we lack care. We make them because we care deeply. In the face of tremendous suffering, prioritization helps us take decisive, thoughtful action instead of freezing or leaving impact on the table. * Surveys show that personal connections are the most common way that people first discover EA. When we share our own stories — explaining not just what we do but why it matters to us emotionally — we help others see that EA offers a concrete way to turn their compassion into meaningful impact. You can also watch my full talk on YouTube. ---------------------------------------- One year ago, I stood on this stage as the new CEO of the Centre for Effective Altruism to talk about the journey effective altruism is on. Among other key messages, my talk made this point: if we want to get to where we want to go, we need to be better at telling our own stories rather than leaving that to critics and commentators. Since
 ·  · 3m read
 · 
A friend of mine who worked as a social worker in a hospital told me a story that stuck with me. She had a conversation with an in-patient having a very difficult time. It was helpful, but as she was leaving, they told her wistfully 'You get to go home'. She found it hard to hear—it felt like an admonition. It was hard not to feel guilt over indeed getting to leave the facility and try to stop thinking about it, when others didn't have that luxury. The story really stuck with me. I resonate with the guilt of being in the fortunate position of being able to go back to my comfortable home and chill with my family while so many beings can't escape the horrible situations they're in, or whose very chance at existence depends on our work. Hearing the story was helpful for dealing with that guilt. Thinking about my friend's situation it was clear why she felt guilty. But also clear that it was absolutely crucial that she did go home. She was only going to be able to keep showing up to work and having useful conversations with people if she allowed herself proper respite. It might be unfair for her patients that she got to take the break they didn't, but it was also very clearly in their best interests for her to do it. Having a clear-cut example like that to think about when feeling guilt over taking time off is useful. But I also find the framing useful beyond the obvious cases. When morality feels all-consuming Effective altruism can sometimes feel all consuming. Any spending decision you make affects how much you can donate. Any activity you choose to do takes time away from work you could be doing to help others. Morality can feel as if it's making claims on even the things which are most important to you, and most personal. Often the narratives with which we push back on such feelings also involve optimisation. We think through how many hours per week we can work without burning out, and how much stress we can handle before it becomes a problem. I do find that