I wrote an essay that's a case study for how Open Philanthropy (and by abstraction, EA) can be better at communications.

It's called Affective Altruism.

I wrote the piece because I was growing increasingly frustrated seeing EA have its public reputation questioned following SBF and OpenAI controversies. My main source of frustration wasn't just seeing EA being interpreted uncharitably, it was that the seeds for this criticism were sewn long before SBF and OpenAI became known entities.

EA's culture of ideological purity and (seemingly intentional) obfuscation from the public sets itself up for backlash. Not only is this unfortunate relative the movement's good intentions, it's strategically unsound. EA fundamentally is in the business of public advocacy. It should be aiming for more than resilience against PR crises. As I say in the piece:

The point of identifying and cultivating a new cause area is not for it to remain a fringe issue that only a small group of insiders care about. The point is that it is paid attention to where it previously wasn't.

The other thing that's frustrating is that what I'm asking for is not for EA to entertain some race-to-the-bottom popularity contest. It's an appeal to respect human psychology, to use time-tested techniques like visualization and story telling that are backed by evidence. There are ways to employ these communications strategies without reintroducing the irrationalities that EA prides itself on avoiding, and without meaningfully diminishing the rigorousness of the movement.

On a final personal note: 

I feel a tremendous love-hate relationship with EA. Amongst my friends (none of which are EAs despite most being inordinately altruistic) I'm slightly embarrassed to call myself an EA. There's a part of me that is allergic to ideologies and in-group dynamics. There's a part of me that's hesitant of allying myself with a movement that's so self-serious and disregarding of outside perceptions. There's also a part of me that feels spiteful towards all the times EA has soft and hard rejected my well-meaning attempts at participation (case-in-point, I've already been rejected from the comms job I wrote this post to support my application for). And yet, I keep coming back to EA because, in a world that is so riddled with despair and confusion, there's something reaffirming about a group of people who want to use evidence to do measurable good. This unimpeachable trait of EA should be understood for the potential energy it wields amongst many people like myself that don't even call themselves EAs. Past any kind of belabored point about 'big tent' movements, all I mean to say is that EA doesn't need to be so closed-off. Just a little bit of communications work would go a long way.

Here's a teaser video I made to go along with the essay:

99

4
1
2

Reactions

4
1
2
Comments19


Sorted by Click to highlight new comments since:

Very useful post, thanks. Improving our comms is one of our three priorities for EA Netherlands in 2024 and this will inform that work. 

Out of interest:

  1. What are your other two priorities?
  2. How will you know if you've been successful in "improving your comms"? Curious to hear if you have a more specific okr here

Hey! 

Our other priorities for 2024 are GCR field building and investing in our volunteering programme. We'll do this alongside maintaining our more established programmes e.g., our national EA crash course, our support for organisers around the country, and our co-working office.

In terms of measuring success, we still need to develop the strategy, so it is not currently possible to say in detail how we will evaluate it. Broadly speaking, we want to increase awareness of, and inclination towards, effective altruism amongst proto-EAs in the Netherlands. We also want to ensure inclination remains high amongst the general public once they become aware of us. Therefore, to evaluate the impact of this work, we will probably conduct surveys to measure awareness and inclination amongst proto-EAs and the general public before and after the interventions outlined in the strategy, whatever they may be, are implemented. 

Of course we'll also keep an eye on basic comms metrics like newsletter subscribers, LinkedIn followers, etc. And downstream metrics like intro programme completions, etc. 

For Q1 our comms OKR is as follows:

Objective: Comms - develop our strategy (ready to be handed to volunteer team) 
 
Key Results  

  1. Get 100 survey responses for our Dutch proto-EA marketing survey by March 8th (this asks about media consumption habits, barriers faced, recommendations for media platforms/influencers, etc).
  2. Internal publication of an analysis of the survey's results by March 15th (we're probably going to miss this deadline, in the end, we decided to rely on a volunteer for the analysis)
  3. Internal publication of a communications strategy (in the style of Rumelt) consisting of a diagnosis, guiding policy, and a set of coherent actions by March 22nd (again, we're probably going to miss this target)
  4. Recruit a team of 3+ volunteers by March 31st to help us implement the strategy (supplementing the marketing strategist and the google ads marketeer we've already got on the team)

Thanks James, cool to hear.

Re your final personal note - I feel a lot like you! Thanks for putting your thoughts out there.

thanks ulrik 🤝

I thought the video was excellent, and the highlights of your article were the concrete ideas and examples of good communication.

More concrete ideas please! I don't think anyone will disagree that EA hasn't been the best at branding itself, but in my experience it's easier said than done!

If people want more concrete ideas they can hire me to communications work.

I don't know how to be more concrete than I did in the article without working for free.

EA's culture of ideological purity and (seemingly intentional) obfuscation from the public sets itself up for backlash.

The link goes to this article itself. Curious what you were trying to link to.

fixed thanks

"An OpenAI program director, who has very little to actually do with this larger public debate, is suddenly subpoenaed to testify in a congressional hearing where they are forced to answer an ill-tempered congress member's questions. It might go something like this"

This should be OP, not OpenAI, right?

fixed thanks

I don't know much about EA yet, so this was nice to hear your perspective about where things could improve. I can see both sides of the coin here; that being accessible helps with distribution of information, and how non-serious people aren't going to include the rigor that's required to understand some of these complex issues. 

I wonder where the middle ground is? I also wonder what changes would bring the most relief to you. Shift in culture? Shift in sharable material?

One of the broader points I'm advocating for is that the middle ground is far more stable and sizable than many in the community might think it is.

I think the 'non-serious' individual you speak of is somewhat of a straw man. If they are real, the risk of them polluting the quality of EA's work is quite small IMO. It's important to make a distinction between the archetype of a follower/fan (external comms) and a worker/creator (internal comms). A lot of EAs conflate internal and external communications.

This is a really cool topic. I wonder why there is tension. I haven't been around long enough to see it in action, but I'm getting a better sense for it as I read similar posts. Do you think there's a key cultural shift that would address the underlying issue? Do you think there's any fear (or some other emotion/rationale) about avoiding this middle ground?

Yeah if you read the essay it spends a lot of time speaking to both of those questions

tldr

The fear is born from the very DNA of EA which has its roots in avoiding emotional irrationalities that lead to ineffective forms of altruism. The culture shift I want to see is a product of a) acknowledging and relinquishing this fear when it's not based on reality b) understanding the value proposition of good communications

The video is interesting! I liked the demonstration at the beginning that you care more about someone's ideas when you have seen who they are. The radio switch at the beginning was a bit long but otherwise very good idea.

Small feedback on your essay itself: 

even as someone interested in hearing what you had to say, your writing could be formatted to let me skim it more efficiently. I'd have loved if you posted more visible TL;DRs at the start & named the sections by their conclusions rather than their guiding questions.

The teaser video worked on me as you predicted though, props on that! 

This also makes for a distinctive cover letter to the OP job, to be sure! Smart.

This is great meta-feedback - I'll be sure to include more TL;DRs at the start of my articles too. 

Curated and popular this week
 ·  · 11m read
 · 
Confidence: Medium, underlying data is patchy and relies on a good amount of guesswork, data work involved a fair amount of vibecoding.  Intro:  Tom Davidson has an excellent post explaining the compute bottleneck objection to the software-only intelligence explosion.[1] The rough idea is that AI research requires two inputs: cognitive labor and research compute. If these two inputs are gross complements, then even if there is recursive self-improvement in the amount of cognitive labor directed towards AI research, this process will fizzle as you get bottlenecked by the amount of research compute.  The compute bottleneck objection to the software-only intelligence explosion crucially relies on compute and cognitive labor being gross complements; however, this fact is not at all obvious. You might think compute and cognitive labor are gross substitutes because more labor can substitute for a higher quantity of experiments via more careful experimental design or selection of experiments. Or you might indeed think they are gross complements because eventually, ideas need to be tested out in compute-intensive, experimental verification.  Ideally, we could use empirical evidence to get some clarity on whether compute and cognitive labor are gross complements; however, the existing empirical evidence is weak. The main empirical estimate that is discussed in Tom's article is Oberfield and Raval (2014), which estimates the elasticity of substitution (the standard measure of whether goods are complements or substitutes) between capital and labor in manufacturing plants. It is not clear how well we can extrapolate from manufacturing to AI research.  In this article, we will try to remedy this by estimating the elasticity of substitution between research compute and cognitive labor in frontier AI firms.  Model  Baseline CES in Compute To understand how we estimate the elasticity of substitution, it will be useful to set up a theoretical model of researching better alg
 ·  · 7m read
 · 
Crossposted from my blog.  When I started this blog in high school, I did not imagine that I would cause The Daily Show to do an episode about shrimp, containing the following dialogue: > Andres: I was working in investment banking. My wife was helping refugees, and I saw how meaningful her work was. And I decided to do the same. > > Ronny: Oh, so you're helping refugees? > > Andres: Well, not quite. I'm helping shrimp. (Would be a crazy rug pull if, in fact, this did not happen and the dialogue was just pulled out of thin air).   But just a few years after my blog was born, some Daily Show producer came across it. They read my essay on shrimp and thought it would make a good daily show episode. Thus, the Daily Show shrimp episode was born.   I especially love that they bring on an EA critic who is expected to criticize shrimp welfare (Ronny primes her with the declaration “fuck these shrimp”) but even she is on board with the shrimp welfare project. Her reaction to the shrimp welfare project is “hey, that’s great!” In the Bible story of Balaam and Balak, Balak King of Moab was peeved at the Israelites. So he tries to get Balaam, a prophet, to curse the Israelites. Balaam isn’t really on board, but he goes along with it. However, when he tries to curse the Israelites, he accidentally ends up blessing them on grounds that “I must do whatever the Lord says.” This was basically what happened on the Daily Show. They tried to curse shrimp welfare, but they actually ended up blessing it! Rumor has it that behind the scenes, Ronny Chieng declared “What have you done to me? I brought you to curse my enemies, but you have done nothing but bless them!” But the EA critic replied “Must I not speak what the Lord puts in my mouth?”   Chieng by the end was on board with shrimp welfare! There’s not a person in the episode who agrees with the failed shrimp torture apologia of Very Failed Substacker Lyman Shrimp. (I choked up a bit at the closing song about shrimp for s
 ·  · 9m read
 · 
Crosspost from my blog.  Content warning: this article will discuss extreme agony. This is deliberate; I think it’s important to get a glimpse of the horror that fills the world and that you can do something about. I think this is one of my most important articles so I’d really appreciate if you could share and restack it! The world is filled with extreme agony. We go through our daily life mostly ignoring its unfathomably shocking dreadfulness because if we didn’t, we could barely focus on anything else. But those going through it cannot ignore it. Imagine that you were placed in a pot of water that was slowly brought to a boil until it boiled you to death. Take a moment to really imagine the scenario as fully as you can. Don’t just acknowledge at an intellectual level that it would be bad—really seriously think about just how bad it would be. Seriously think about how much you’d give up to stop it from happening. Or perhaps imagine some other scenario where you experience unfathomable pain. Imagine having your hand taped to a frying pan, which is then placed over a flame. The frying pan slowly heats up until the pain is unbearable, and for minutes you must endure it. Vividly imagine just how awful it would be to be in this scenario—just how much you’d give up to avoid it, how much you’d give to be able to pull your hand away. I don’t know exactly how many months or years of happy life I’d give up to avoid a scenario like this, but potentially quite a lot. One of the insights that I find to be most important in thinking about the world is just how bad extreme suffering is. I got this insight drilled into me by reading negative utilitarian blogs in high school. Seriously reflecting on just how bad extreme suffering is—how its intensity seems infinite to those experiencing it—should influence your judgments about a lot of things. Because the world is filled with extreme suffering. Many humans have been the victims of extreme suffering. Throughout history, tort