Slightly independent to the point Habryka is making, which may well also be true, my anecdotal impression is that the online EA community / EAs I know IRL were much bigger on 'we need to beat China' arguments 2-4 years ago. If so, simple lag can also be part of the story here. In particular I think it was the mainstream position just before ChatGPT was released, and partly as a result I doubt an 'overwhelming majority of EAs involved in AI safety' disagree with it even now.
Example from August 2022:
https://www.astralcodexten.com/p/why-not-slow-ai-progress
So maybe (the argument goes) we should take a cue from the environmental activists, and be hostile towards AI companies...
This is the most common question I get on AI safety posts: why isn’t the rationalist / EA / AI safety movement doing this more? It’s a great question, and it’s one that the movement asks itself a lot...
Still, most people aren’t doing this. Why not?
Later, talking about why attempting a regulatory approach to avoiding a race is futile:
The biggest problem is China. US regulations don’t affect China. China says that AI leadership is a cornerstone of their national security - both as a massive boon to their surveillance state, and because it would boost their national pride if they could beat America in something so cutting-edge.
So the real question is: which would we prefer? OpenAI gets superintelligence in 2040? Or Facebook gets superintelligence in 2044? Or China gets superintelligence in 2048?
Might we be able to strike an agreement with China on AI, much as countries have previously made arms control or climate change agreements? This is . . . not technically prevented by the laws of physics, but it sounds really hard. When I bring this challenge up with AI policy people, they ask “Harder than the technical AI alignment problem?” Okay, fine, you win this one.
I feel like a generic non-EA policy person reading that post could well end up where the congressional commission landed? It's right there in the section that most explicitly talks about policy.
For the record, I have a few places I think EA is burning >$30m per year, not that AW is actually one of them. Most EAs I speak to seem to have similarly-sized bugbears? Though unsurprisingly they don't agree about where the money is getting burned..
So from where I stand I don't recognise your guess of how people respond to that situation. A few things I believe that might help explain the difference:
I don't expect to have capacity to engage further here, but if further discussion suggests that one of the above is a particularly surprising claim, I may consider writing it up in more detail in future.
If you weigh desires/preferences by attention or their effects on attention (e.g. motivational salience), then the fact that intense suffering is so disruptive and would take priority over attention to other things in your life means it would matter a lot.
Recently I failed to complete a dental procedure because I kept flinching whenever the dentist hit a particularly sensitive spot. They needed me to stay still. I promise you I would have preferred to stay still, not least because what ended up happening was I had to have it redone and endured more pain overall. My forebrain understood this, my hindbrain is dumb.
(FWIW the dentist was very understanding, and apologetic that the anesthetic didn't do its job. I did not get the impression that my failure was unusual given that.)
When I talk about suffering disrupting enjoyment of non-hedonic goods I mean something like that flinch; a forced 'eliminate the pain!' response that likely made good sense back in the ancestral environment, but not a choice or preference in the usual sense of that term. This is particularly easy to see in cases like my flinch where the hindbrain's 'preference' is self-defeating, but I would make similar observations in some other cases, e.g. addiction.
If you don't weigh desires by attention or their effects on attention, I don't see how you can ground interpersonal utility comparisons at all
I don't quite see what you're driving at with this line of argument.
I can see how being able to firmly 'ground' things is a nice/helpful property for an theory of 'what is good?' to have. I like being able to quantify things too. But to imply that measuring good must be this way seems like a case of succumbing to the Streetlight Effect, or perhaps even the McNamara fallacy if you then downgrade other conceptions of good in the style of below quote.
Put another way, it seems like you prefer to weight by attention because it makes answers easier to find, but what if such answers are just difficult to find?
The fact that 'what is good?' has been debated for literally millenia with no resolution in sight suggests to me that it just is difficult to find, in the same way that after some amount of time you should acknowledge your keys just aren't under the streetlight.
But when the McNamara discipline is applied too literally, the first step is to measure whatever can be easily measured. The second step is to disregard that which can't easily be measured or given a quantitative value. The third step is to presume that what can't be measured easily really isn't important. The fourth step is to say that what can't be easily measured really doesn't exist.
To avoid the above pitfall, which I think all STEM types should keep in mind, when I suspect my numbers are failing to capture the (morally) important things my default response is to revert in the direction of Common sense (morality). I think STEM people who fail to check themselves this way often end up causing serious harm[1]. In this case that would make me less inclined to trade human lives for aminal welfare, not more.
I'll probably leave this post at this point unless I see a pressing need for further clarification of my views. I do appreciate you taking the time to engage politely.
Hi Michael,
Sorry for putting off responding to this. I wrote this post quickly on a Sunday night, so naturally work got in the way of spending the time to put this together. Also, I just expect people to get very upset with me here regardless of what I say, which I understand - from their point of view I'm potentially causing a lot of harm - but naturally causes procrastination.
I still don't have a comprehensive response, but I think there are now a few things I can flag for where I'm diverging here. I found titotal's post helpful for establishing the starting point under hedonism:
For the intervention of cage free campaigns, using RP's moral weights, the intervention saves 1996 DALY's per thousand dollars, about a 100 times as effective as AMF.
However, even before we get into moral uncertainty I think this still overstates the case:
Animal welfare (AW) interventions are much less robust than the Global Health and Development (GHD) interventions animal welfare advocates tend to compare them to. Most of them are fundamentally advocacy interventions, which I think advocates tend to overrate heavily.
How to deal with such uncertainty has been the topic of much debate, which I can't do justice to here. But one thing I try to do is compare apples-to-apples for robustness where possible; if I relax my standards for robustness and look at advocacy, how much more estimated cost-effectiveness do I get in the GHD space? Conveniently, I currently donate to Giving What We Can as 'Effective Giving Advocacy' and have looked into their forward-looking marginal multiplier a fair bit; I think it's about 10x. Joel Tan looked and concluded 13x. I've checked with others who have looked at GWWC in detail; they're also around there. I've also seen 5x-20x claims for things like lead elimination advocacy, but I haven't looked into those claims in nearly as much detail.
Overall I think that if you're comfortable donating to animal welfare interventions, comparing to AMF/Givewell 'Top Charities' is just a mistake; you should be comparing to the actual best GHD interventions under your tolerance for shaky evidence, which will have estimated cost-effectiveness 10x higher or possibly even more.
Also, I subjectively feel like AW is quite a bit less robust than even GHD advocacy; there's a robustness issue from advocacy in both cases, but AW also really struggles with a lack of feedback loops - we can't ask the animals how they feel - and so I think is much more likely to end up causing harm on its own terms. I don't know how to quantify this issue, and it doesn't seem like a huge issue for cage-free specifically, so will set this aside. Back when AW interventions were more about trying to end factory farming rather than improving conditions on factory farms it did worry me quite a bit.
After the two issues I am willing to quantify we're down to around 3.3x, and we're still assuming hedonism.
On the other hand, I have the impression that RP made an admirable effort to tend towards conservatism in some empirical assumptions, if not moral ones. I think Open Phil also tends this way sometimes. So I'm not as sure as I usually would what happens if somebody looks more deeply; overwhelmingly I would say EA has found that interventions get worse the more you look at them, which is a lot of why I penalise non-robustness in the first place, but perhaps Open Phil + RP have been conservative enough that this isn't the case?
***
Still, my overall guess is that if you assume hedonism AW comes out ahead. I am not a moral realist; if people want to go all-in on hedonism and donate to AW on those grounds, I don't see that I have any grounds to argue with them. But as my OP alluded to, I tend to think there is more at stake / humans are 'worth more' in the non-hedonic worlds. So when I work through this I end up underwhelmed by the overall case.
***
This brings us to the much thornier territory of moral uncertainty. While continuing to observe that I'm out of my depth philosophically, and am correspondingly uncertain how best to approach this, some notes on how I think about this and where I seem to be differing:
I find experience machine thought experiments, and people's lack of enthusiasm for them, much more compelling than 'tortured Tim' thought experiements for trying to get a handle on how much of what matters is pleasure/suffering. The issue I see with modelling extreme suffering is that it tends to heavily disrupt non-hedonic goods, and so it's hard to figure out how much of the badness is the suffering versus the disruption. We can get a sense of how much people care about this disruption from their refusal to enter the experience machine; a lot of the rejections I see and personally feel boil down to "I'm maxing out pleasure but losing everything that 'actually matters'".
RP did mention this but I found their handling unconvincing; they seem to have very different intuitions to me for how much torture compromises human ability to experience what 'actually matters'. Empirical evidence from people with chronic nerve damage is similarly tainted by the fact that e.g. friends often abandon you when you're chronically in pain, you may have to drop hobbies that meant a lot to you, and so on.
I've been lucky enough never to experience anything that severe, but if I look at the worst periods of my life it certainly seemed like a lot more impact came from these 'secondary' effects - interference with non-hedonic goods - than from the primary suffering. My heart goes out to people who are dealing with worse conditions and very likely taking larger 'secondary' hits.
I also just felt like the Tortured Tim thought experiment didn't 'land' even on its own terms for me, similar to the sentiments expressed in this comment and this comment.
Ah, gotcha, I guess that works. No, I don't have anything I would consider strong evidence, I just know it's come up more than anything else in my few dozen conversations over the years. I suppose I assumed it was coming up for others as well.
they should definitely post these and potentially redirect a great deal of altruistic funding towards global health
FWIW this seems wrong, not least because as was correctly pointed out many times there just isn't a lot of money in the AW space. I'm pretty sure GHD has far better places to fundraise from.
To the extent I have spoken to people (not Jeff, and not that much) about why they don't engage more on this, I thought the two comments I linked to in my last comment had a lot of overlap with the responses.
I think there has been very little argument for animals not counting at the post level because, quite simply, the argument that (in expectation) they do count is just far stronger.
I'm confused how this works, could you elaborate?
My usual causal chain linking these would be 'argument is weak' -> '~nobody believes it' -> 'nobody posts it'.
The middle step fails here. Do you have something else in mind?
FWIW, I thought these two comments were reasonable guesses at what may be going on here.
First, want to flag that what I said was at the post level and then defined stronger as:
the points I'm most likely to hear and give most weight to when discussing this with longtime EAs in person
You said:
I haven’t heard any arguments I’d consider strong for prioritizing global health which weren’t mentioned during Debate Week
So I can give examples of what I was referring to, but to be clear we're talking somewhat at cross purposes here:
With that in mind, I would say that the most common argument I hear from longtime EAs is variants of 'animals don't count at all'. Sometimes it's framed as 'almost certainly aren't sentient' or 'count as ~nothing compared to a child's life'. You can see this from Jeff Kaufman, and Eliezer Yudkowsky, and it's one I hear a decent amount from EAs closer to me as well.
If you've discussed this a ton I assume you have heard this too, and just aren't thinking of the things people say here as strong arguments? Which is fine and all, I'm not trying to argue from authority, at least not at this time. My intended observation was 'lots of EAs think a thing that is highly relevant to the debate week, none of them wrote it up for the debate week'.
I think that observation holds, though if you still disagree I'm curious why.
Thanks for this post, I was also struggling with how scattered the numbers seemed to be despite many shared assumptions. One thing I would add:
Another thing I want to emphasise: this is an estimate of past performance of the entire animal rights movement. It is not an estimate of the future cost effectiveness of campaigns done by EA in particular. They are not accounting for tractableness, neglectedness, etc of future donations....
In the RP report, they accounted for this probable drop in effectiveness by dropping the effectiveness by a range of 20%-60%. This number is not backed up by any source: it appears to be a guesstimate based on the pros and cons listed by Saulius. Hence there is signifcant room for disagreement here.
If we take Sallius’s estimate of 42 chickens affected per dollar, and discount it by 40%, we get a median of 42*0.6 = 25.2 chickens affected per dollar.
Last year Open Phil stated that their forward-looking estimate was 5x lower than Saulius' backward-looking estimate. This is the type of 'narrow' question I default defer to Open Phil on, and so I would drop your final figures by a factor of 3, going from 60% of Saulius to 20% of Saulius.
That sounds plausible. I do think of ACX as much more 'accelerationist' than the doomer circles, for lack of a better term. Here's a more recent post from October 2023 informing that impression, below probably does a better job than I can do of adding nuance to Scott's position.
https://www.astralcodexten.com/p/pause-for-thought-the-ai-pause-debate