AI safety researcher
Ok interesting! I'd be interested in seeing this mapped out a bit more, because it does sound weird to have BOS be offsettable with positive wellbeing, positive wellbeing to be not offsettable with NOS, but BOS and NOS are offsetable with each other? Or maybe this isn't your claim and I'm misunderstanding
This is what kills the proposal IMO, and EJT also pointed this out. The key difference between this proposal and standard utilitarianism where anything is offsettable isn't the claim that that NOS is worse than TREE(3) or even 10^100 happy lives, since this isn't a physically plausible tradeoff we will face anyway. It's that once you believe in NOS, transitivity compels you to believe it is worse than any amounts of BOS, even a variety of BOS that, according to your best instruments, only differs from NOS in the tenth decimal place. Then once you believe this, the fact that you use a utility function compels you to create arbitrary amounts of BOS to avoid a tiny probability of a tiny amount of NOS.
It is not necessary to be permanently vegan for this. I have only avoided chicken for about 4 years, and hit all of these benefits.
I separately believe that social and political change are pretty small compared to EA animal welfare efforts. But beef and high-welfare-certified meat options cut down on suffering by >90% vs factory farmed chicken (or eggs, squid, and some fish) and also serve many of the signaling benefits. If you eat welfare-certified animal products only, it may even be higher for two reasons:
I perceive it as +EV to me but I feel like I'm not the best buyer of short timelines. I would maybe do even odds on before 2045 for smaller amounts, which is still good for you if you think the yearly chance won't increase much. Otherwise maybe you should seek a bet with someone like Eli Lifland. The reason I'm not inclined to make large bets is that the markets would probably give better odds for something that unlikely, eg options that pay out with very high real interest rates; whereas a few hundred dollars is enough to generate good EA forum discussion.
No bet. I don't have a strong view on short timelines or unemployment. We may find a bet about something else; here are some beliefs
A footnote says the 0.15% number isn't an actual forecast: "Participants were asked to indicate their intuitive impression of this risk, rather than develop a detailed forecast". But superforecasters' other forecasts are roughly consistent with 0.15% for extinction, so it still bears explaining.
In general I think superforecasters tend to anchor on historical trends, while AI safety people anchor on what's physically possible or conceivable. Superforecasters get good accuracy compared to domain experts on most questions because domain experts in many fields don't know how to use reference classes and historical trends well. But it's done poorly recently because progress has accelerated-- even in 2022 superforecasters' median for the AI IMO gold medal was 2035, whereas it actually happened in 2025. Choosing a reference class for extinction is very difficult so people just rely on vibes.
Let's take the question of whether world energy consumption will double year-over-year before 2045. In the full writeup, superforecasters, whose median is 0.35%, emphasized the huge difficulty in constructing terrestrial facilities to use that much energy:
Superforecasters generally expressed skepticism about a massive increase (doubling) in global energy consumption, due to this having low base rates and requiring unlikely technical breakthroughs.
- Many rationales expressed skepticism that the rate of energy production could be scaled up so quickly even with advanced AI.
- The breakthroughs thought to be needed are in energy production and distribution techniques.
- A few superforecasters said that they thought fusion was the only plausible path, but even then other physical infrastructure might be limiting.
In contrast, I wrote about how doubling energy production in a year starting from self replicating robots in space just requires us to be more than ~0.1% efficient in refining asteroid raw material into solar panels and robots, and that it's likely we get there eventually. I'm closer to 50% on this question.
Dyson swarms can have energy doubling times of *days*. The energy payback time of current solar panels on Earth is 1-2 years, in space there's 8x more light than on Earth, and we're >3 OOMs away from the minimum energy required to make solar panels (reducing SiO2 to Si).
I think to *not* get an energy doubling in one year by the time we exhaust the solar system's energy, it would require a big slowdown (eg due to regulation or low energy demand) through about 15 OOMs of energy use, spanning from the first decently efficient self-replicating robots through Dyson swarms until we disassemble the gas planets for fusion fuel. Such a period would necessarily take decades or centuries to always be doubling slower than 1 year, which is basically an eternity when we have ASI.
The other factor is that AI safety people sometimes have a more inclusive definition of p(doom), that includes not just extinction but AIs seizing control of the world and colonizing the galaxy while leaving humans powerless.
I think I would take your side here. Unemployment above 8% requires replacing so many jobs that humans can't find new ones elsewhere even during the economic upswing created by AGI, and there is less than 2 years until the middle of 2027. This is not enough time for robotics (on current trends robotics time horizons will be under 1 hour) and AI companies can afford to keep hiring humans even if they wouldn't generate enough value most places, so the question is whether we see extremely high unemployment in remotable sectors that automate away existing jobs but don't have huge labor productivity gains from AI. 2029 could be a different story.
Would like to see David's perspective here, whether he just has short timelines or has some economic argument too.
Spreading around the term “humane meat” may get it into some people’s heads that this practice can be humane, which could in turn increase consumption overall, and effectively cancel out whatever benefits you’re speculating about.
I don't know what the correct definition of "humane" is, but I strongly disagree with this claim in the second half. The question is whether higher-welfare imports reduce total suffering once we account for demand effects. So we should care about improving conditions from "torture camps" -> "prisons" -> "decent". Torture camps are many times worse than merely "inhumane"!
The average consumer (who eats a ton of super high suffering chicken, doesn't know that most chicken is torture chicken, and doesn't believe labels anyway) wouldn't eat much more chicken overall when the expensive chicken with the non-fraudulent "humane" label lowers in price. Nor would enough vegetarians start eating chicken because they're only 5% of the US population and many of those are motivated by religion or health.
More likely, there will need to be a huge effort to get consumers to understand that they should spend anything on lower-suffering chicken, then another to get grocers to not mark up the price anyway, after which implementing this policy could replace 260 million torture camp chicken lives with maybe 300 million slightly uncomfortable chicken lives. (With a net increase mostly due to competition lowering the price of higher-suffering chicken.)
One can object to actually implementing this policy on deontological or practical grounds, but on consequences, high-suffering chicken is many times worse than "inhumane" pasture-raised chicken, so the demand increase would not even be close to canceling out the benefits unless you have a moral view under which everything inhumane is equally bad. I wish we were in a world where we could demand that food be 100% humane, but ignoring the principle of triage is why EA animal advocates, not purity-focused ones, have prevented billions of years of torture.
Yeah, because I believe in EA and not in the socialist revolution, I must believe that EA could win some objective contest of ideas over socialism. In the particular contest of EA -> socialist vs socialist -> EA conversions I do think EA would win since it's had a higher growth rate in the period both existed, though someone would have to check how many EA deconverts from the FTX scandal became socialists. This would be from both signal and noise factors; here's my wild guess at the most important factors:
But I think someone would actually need to do that experiment or at least gather the data
There are a few mistakes/gaps in the quantitative claims:
This is not quite the same as either property 3 or property 3' in the Wikipedia article, and it's plausible but unclear to me that you can prove 3' from it. Property 3 uses "p ∈ [0, 1]" and 3' has an inequality; it seems like the argument still goes through with 3' so I'd switch to that, but then you should also say why 3 is unintuitive to you because VNM only requires 3 OR 3'.
"Grow without bound" just means that for any M, we have f(X) > M for sufficiently large X. This is different from there being a vertical asymptote so a threshold is not inevitable. For instance one could have f(X) = X or f(X) = X^2.
It would be confusing to call this behavior continuous, because (a) the VNM axiom you reject is called continuity and (b) we are not using any other properties of the extended reals, but we are using real-valued probabilities and x values.
This may seem like a nitpick, but "write down", "compute", and "physically instantiate" are wildly different ranges of numbers. The largest number one could "physically instantiate" is something like 10^50 minds, the most one could "write down" the digits of is something like 10^10^10.
Not all large numbers are the same here, because if one thinks the offset ratio for a cluster headache is in the 10^50 range, there are only 50 'levels' of suffering each of which is 10x worse than the last. If it's over 10^10^10, there are over 10 billion such 'levels', it would be impossible to rate cluster headaches on a logarithmic pain scale, and we would happily give everyone on Earth (say) a level 10,000,000,000 cluster headache to prevent one person from having a (slightly worse than average) level 10,000,000,010 cluster headache. Moving from 10^10^10 to infinity, we would then believe that suffering has a threshold t where t + epsilon intensity suffering cannot be offset by removing t - epsilon intensity suffering, and also need to propose some other mechanism like lexicographic order for how to deal with suffering above the infinite badness threshold.
So it's already a huge step to reject numbers we can "physically instantiate" to ones we can barely "write down", and another step from there to infinity; at both steps your treatment of comparisons between different suffering intensities changes significantly, even in thought experiments without an unphysically large number of beings.