Hide table of contents

Risk neutrality is the idea that a 50% chance of $2 is equally valuable as a 100% chance of $1. This is a highly unusual preference, but it sometimes makes sense:

Exampleville is infested with an always-fatal parasite. All 1,024 residents have the worm.

For each gold coin you pay, a pharmacist will deworm one randomly selected resident.

You only have one gold coin, but the pharmacist offers you a bet. Flip a weighted coin, which comes up heads 51% of the time. If heads, you double your money. If tails, you lose it all.

You can play as many times as you like.

It makes sense to be risk neutral here. If you keep playing until you've saved everyone or go broke, each resident gets a 0.51^10 = 1 in 840 chance of survival. If you don't play, it's 1 in 1,024.

For a few of the most scalable interventions, I think EA is like this. OpenPhil seems to expect a very low rate of diminishing returns for GiveWell's top charities. I wouldn't recommend a double-or-nothing on GiveWell's annual budget at a 51% win rate, but I would at 60%.[1] 

Most interventions aren't like this. The second $100M of AI safety research buys much less than the first $100M. Diminishing returns seem really steep. 

SBF and Caroline Ellison had much more aggressive ideas. More below on that.

Intro to the Kelly Criterion

Suppose you're offered a 60% double-or-nothing bet. How much should you wager? In 1956, John Kelly, a researcher at Bell Labs, proposed a general solution. Instead of maximizing Expected Value, he maximized the geometric rate of return. Over a very large number of rounds, this is given by

Where  is the proportion of your bankroll wagered each round, and  is the probability of winning each double-or-nothing bet.

Maximizing with respect to  gives

Therefore, if you have a 60% chance of winning, bet 20% of your bankroll on each round. At a 51% chance of winning, bet 2%.

This betting method, the Kelly Criterion, has been highly successful in real life. The MIT Blackjack Team bet this way, among many others.

Implicit assumption in the Kelly Criterion

The Kelly Criterion sneaks in the assumption that you want to maximize the geometric rate of return. 

This is normally a good assumption. Maximizing the geometric rate of return makes your bankroll approach infinity as quickly as possible!

But there are plenty of cases there the geometric rate of return is a terrible proxy for your underlying preferences. In the Parasite example above, your bets have a negative geometric rate of return, but they're still a good decision.

Most people also have a preference for reducing the risk of losing much of their starting capital. Betting a fraction of the Kelly Bet gives a slightly lower rate of return but a massively lower risk of ruin. 

Above: distribution of outcomes after 100 rounds with a 60% chance of winning each bet. The red line (Full Kelly) wagers 20% of the bankroll on each round, while the blue (.4 Kelly) wagers 8%. The x-axis uses a log scale: "2" means multiplying your money a hundredfold, while "-1" means losing 90% of your money.

Crucially, neither of these distributions is better. The red gets higher returns, but the blue has a much more favorable risk profile. Practically, 0.2 to 0.5 Kelly seems to satisfy most peoples' real-world preferences.

Treat risk preferences like any other preference

SBF and Caroline Ellison both endorsed full Martingale strategies. Caroline says:

Most people don’t do this, because their utility is more like a function of their log wealth or something and they really don’t want to lose all of their money. Of course those people are lame and not EAs; this blog endorses double-or-nothing coin flips and high leverage.

She doesn't specify that it's a weighted coin, but I assume she meant a positive EV bet. 

Meanwhile, SBF literally endorses this strategy for the entire Earth in an interview with Tyler Cowen

[TYLER] ... 51 percent, you double the Earth out somewhere else; 49 percent, it all disappears. Would you play that game? And would you keep on playing that, double or nothing?

SBF agrees to play the game repeatedly, with the caveats that the new Earths are created in noninteracting universes, that we have perfect confidence that the new Earths are in fact created, and that the game is played a finite number of times.

I strongly disagree with SBF. But it's a sincerely held preference, not a mistake about the underlying physical reality. If the game is played for 10 rounds, the choices are

A. One Earth, as it exists today
B. 839/840 chance that Earth is destroyed and a 1/840 chance of 1,024 Earths

I strongly prefer A, and SBF seems to genuinely prefer B

One of my objections to B is that I don't believe a life created entirely makes up for a life destroyed, and I don't believe that two Earths, even in noninteracting universes, is fully twice as valuable as one. I also think population ethics is very unsolved, so I'd err toward caution.

That's a weak objection though. Tyler could keep making the odds more favorable until he's overcome each of these issues. Here's Tyler's scenario on hard mode:

50%, you create a million Earths in noninteracting universes. 50 percent, it all disappears. Would you play that game? For how many rounds? Each round multiplies the set of Earths a millionfold.

I still wouldn't play. My next objection is that it's terrible for SBF to make this decision for everyone, when I'm confident almost the entire world would disagree. He's imposing his preferences on the rest of the world, and it'll probably kill them. That sounds like the definition of evil.

That's still a weak objection. It doesn't work in the least convenient world, because Tyler could reframe the scenario again:

50%, you create a million Earths in noninteracting universes. 50 percent, it all disappears. In a planetwide election, 60% of people agreed to play. You are a senior official with veto power. If you don't exercise your veto, then Earth will take the risk.

I won't answer this one. I'm out of objections. I hope I'd do the right thing, but I'm not sure what that is.

Did these ideas affect the FTX crash?

I think so. I predict with 30% confidence that, by end of 2027, a public report on the FTX collapse by a US Government agency will identify SBF or Caroline Ellison's attitude towards risk as a necessary cause of the FTX crash and will specifically cite their writings on double-or-nothing bets.

By "necessary cause" I mean that, if SBF and Caroline did not have these beliefs, then the crash would not have occurred. For my prediction, it's sufficient for the report to imply this causality. It doesn't have to be stated directly.

I have about 80% confidence in the underlying reality (risk attitude was a necessary cause of the crash), but that's not a verifiable prediction.

Of course, it's not the only necessary cause. The fraudulent loans to Alameda are certainly another one!

Conclusions

A fractional Kelly Bet gets excellent results for the vast majority of risk preferences. I endorse using 0.2 to 0.5 Kelly to calibrate personal risk-taking in your life.

I don't endorse risk neutrality for the very largest EA orgs/donors. I think this cashes out as a factual disagreement with SBF. I'm probably just estimating larger diminishing returns than he is.

For everyone else, I literally endorse wagering 100% of your funds to be donated on a bet that, after accounting for tax implications, amounts to a 51% double-or-nothing. Obviously, not in situations with big negative externalities.

I disagree strongly with SBF's answer to Tyler Cowen's Earth doubling problem, but I don't know what to do once the easy escapes are closed off. I'm curious if someone does.

Cross-posted to LessWrong

  1. ^

    Risk neutral preferences didn't really affect the decision here. We would've gotten to the same decision (yes on 60%, no on 51%) using the Kelly Criterion on the total assets promised to EA

Comments6


Sorted by Click to highlight new comments since:

In Exampleville (parasite example), if money is infinitely divisible, Kelly betting will save everyone with probability 1.

If bets must be an integral number of coins (and you start with only one coin, a very inconvenient world!),  the strategy of "bet the highest amount not more than Kelly, except always bet at least one coin and never bet more than necessary to reach 1024", will save everyone with probability of approximately 0.038, or about 1 in 26, which is significantly better than 1 in 840.

In an intermediate scenario where gold coins are divisible into 100 smaller units for betting purposes, the probability for an analogous strategy will exceed 0.94.

While these strategies do involve betting higher than Kelly in the sole situation that you bet your last betting unit instead of giving up, in general it is fair to call them Kelly variants in contrast to "always bet everything".

The reason that Kelly variants are vastly superior to always betting everything is that the terms of your hypothetical are that "You can play as many times as you like." It is difficult to conceive of why anyone would risk more than Kelly with this allowance (again with exceptions for your last betting unit).

You can salvage the example by saying that you are limited to 10 bets.

However, if you are limited to, for example, 20 bets, then the best strategy will be some intermediate that is riskier than Kelly but definitely less risky than "always bet everything".

My advice to readers is that betting more than Kelly is usually a terrible idea and if you think it's a good idea, you're probably making a mistake. The fact that a thought experiment, in an intelligently written article, went awry and gave an example where risk neutral betting is worse than Kelly variants when it was supposed to do the opposite, is a great example of the danger.

Nonetheless, I concede that for small and medium donors, betting 100% of your funds-to-be-donated on a 51% coin flip is fine if it is genuinely limited to your personal donation. However, the reason for this concession is not that I endorse risk-neutral betting for EA, but rather that the correct way to look at the bet in this scenario, when the funds are already "to be donated", is not as part of your personal bankroll, but rather as part of the bankroll of the cause for which you intend to donate them. Since any individual small or medium donor will in all likelihood represent less than 0.1% of the bankroll of the cause, such an individual betting such funds in isolation is in practice never exceeding Kelly betting.

I would still caution against 51% coin flips in a scenario where a large fraction of the EA community coordinates, spreads news of "an opportunity", etc., such that a large fraction of EA funds are risked on highly correlated 51% coin flips, for reasons which are hopefully quite obvious.

I read the example as one in which the pharmacist's only offer is for a series of double-or-lose-all wagers. That is unlikely to be a real-world scenario, admittedly.

Maybe a more realistic hypo would allow betting in tenths of a coin, but after each round, some small fraction of the population is killed by the parasite. Very conservative risk management always looks good if you have a practically infinite time to flip coins before you need to act.

I had the same reading, but then the thought experiment doesn't do anything to justify why double or nothing is better than Kelly betting. It's just better than... not doing anything, I guess, which isn't an impressive conclusion.

I think it's moderately effective as a simplified model that quickly demonstrates the premise in the title could be true in some scenarios.

The basic Kelly bet hypo, as I understand it, also contains some assumptions that are uncommon in the real world (indefinite number of available rounds that result instantly). It's a closer match to many real-world scenarios, but I'd be open to an argument that there are practical scenarios are closer to OP's simplified model than the basic Kelly model. That's what I would be interested in hearing more from OP about.

Thanks, it didn't occur to me that "double your money" in the example meant "double your entire bankroll" not "double some amount that you selected to bet". I think you're right that that's what the author meant, but I share Karthik's concern that you're then comparing "bet everything" to "do nothing".

I understand your concerns about the number of coin flips required. I loosely addressed it in my original comment, noting that even with a fixed limit of 20 bets in Exampleville (but with the ability to select bet size), the optimal approach will be clearly be less risky than "always bet everything".

Furthermore, it's important to consider that any scenario based on a limited number of bets is only applicable to the real world if the bets in the scenario will be the only positive-EV bets you can make in your entire life, or if the resource you are betting will never have any utility outside of the constrained scenario, or something very strange like that. In the real world, if you encounter a positive-EV bet denominated in USD that you are only able to make a limited number of times, even if you have a salient group of people to aid with the outcome of the bet, you need to consider that you may have other positive-EV bets in your life and you may be able to use the USD to help other people in the future if you manage risk more normally and don't blow your entire net worth on the first set of positive-EV bets that you encounter.

With respect, it is highly unusual to call full Kelly betting "very conservative risk management". Most people familiar with the concept consider full Kelly betting to be extremely aggressive.

I definitely wouldn't call full Kelly "very conservative." My point was that in Exampleville and other scenarios with a capped best outcome, a very conservative strategy (e.g., 0.001 Kelly) is trivially superior to even merely conservative strategies if there is practically infinite time to flip coins before you need to act, and flipping coins is a costless activity. So the specification of the constraints is what makes the problem interesting (and realistic). The example shows that there are constraints where very agressive betting makes sense; the rest of the post needs to demonstrate how those scenarios -- or at least somewhat similar scenarios -- might occur in real life.

Bets don't have to be in USD or anything else that is a store of value. One could imagine that the "currency" was virologists in a rapidly-moving pandemic. You can put them on safe research that will save a modest number of lives or tell them to roll the dice on speculative research with a possible big payoff. You can't really "store" their labor for a possible better bet that may appear ten years down the road, or use it for something else very effectively.

Curated and popular this week
 ·  · 32m read
 · 
Summary Immediate skin-to-skin contact (SSC) between mothers and newborns and early initiation of breastfeeding (EIBF) may play a significant and underappreciated role in reducing neonatal mortality. These practices are distinct in important ways from more broadly recognized (and clearly impactful) interventions like kangaroo care and exclusive breastfeeding, and they are recommended for both preterm and full-term infants. A large evidence base indicates that immediate SSC and EIBF substantially reduce neonatal mortality. Many randomized trials show that immediate SSC promotes EIBF, reduces episodes of low blood sugar, improves temperature regulation, and promotes cardiac and respiratory stability. All of these effects are linked to lower mortality, and the biological pathways between immediate SSC, EIBF, and reduced mortality are compelling. A meta-analysis of large observational studies found a 25% lower risk of mortality in infants who began breastfeeding within one hour of birth compared to initiation after one hour. These practices are attractive targets for intervention, and promoting them is effective. Immediate SSC and EIBF require no commodities, are under the direct influence of birth attendants, are time-bound to the first hour after birth, are consistent with international guidelines, and are appropriate for universal promotion. Their adoption is often low, but ceilings are demonstrably high: many low-and middle-income countries (LMICs) have rates of EIBF less than 30%, yet several have rates over 70%. Multiple studies find that health worker training and quality improvement activities dramatically increase rates of immediate SSC and EIBF. There do not appear to be any major actors focused specifically on promotion of universal immediate SSC and EIBF. By contrast, general breastfeeding promotion and essential newborn care training programs are relatively common. More research on cost-effectiveness is needed, but it appears promising. Limited existing
 ·  · 2m read
 · 
Summary: The NAO will increase our sequencing significantly over the next few months, funded by a $3M grant from Open Philanthropy. This will allow us to scale our pilot early-warning system to where we could flag many engineered pathogens early enough to mitigate their worst impacts, and also generate large amounts of data to develop, tune, and evaluate our detection systems. One of the biological threats the NAO is most concerned with is a 'stealth' pathogen, such as a virus with the profile of a faster-spreading HIV. This could cause a devastating pandemic, and early detection would be critical to mitigate the worst impacts. If such a pathogen were to spread, however, we wouldn't be able to monitor it with traditional approaches because we wouldn't know what to look for. Instead, we have invested in metagenomic sequencing for pathogen-agnostic detection. This doesn't require deciding what sequences to look for up front: you sequence the nucleic acids (RNA and DNA) and analyze them computationally for signs of novel pathogens. We've primarily focused on wastewater because it has such broad population coverage: a city in a cup of sewage. On the other hand, wastewater is difficult because the fraction of nucleic acids that come from any given virus is very low,[1] and so you need quite deep sequencing to find something. Fortunately, sequencing has continued to come down in price, to under $1k per billion read pairs. This is an impressive reduction, 1/8 of what we estimated two years ago when we first attempted to model the cost-effectiveness of detection, and it makes methods that rely on very deep sequencing practical. Over the past year, in collaboration with our partners at the University of Missouri (MU) and the University of California, Irvine (UCI), we started to sequence in earnest: We believe this represents the majority of metagenomic wastewater sequencing produced in the world to date, and it's an incredibly rich dataset. It has allowed us to develop
GiveWell
 ·  · 2m read
 · 
Recent cuts to US government foreign assistance have destabilized global health programs, impacting some of the most cost-effective interventions we’ve found for saving and improving lives, such as malaria nets, malaria chemoprevention, and community-based management of acute malnutrition. This situation is a major focus of our research team at the moment, and we’re working to balance a targeted, near-term response to urgent needs with a broad, long-term perspective of needs that may emerge. The US has historically provided roughly 20% to 25% ($12 billion to $15 billion) of the total global aid to support health programs, which combat malaria, tuberculosis, HIV, maternal and child health issues, and much more.[1] While the long-term effects remain uncertain and exact numbers remain difficult to ascertain, cuts of 35% to 90% of US foreign aid dollars are being publicly discussed by the administration.[2] We’ve created a webpage to provide an overview of how we’re responding, and we’ve started to record a series of conversations with our research team that shares timely snapshots of this rapidly evolving situation. Our first episode shared a broad overview of the impacts of US government aid cuts and GiveWell’s initial response. In our newly released second episode, GiveWell Program Officer Natalie Crispin joins CEO and co-founder Elie Hassenfeld to zoom in on a specific case, focusing on grants we’ve made to support urgent funding gaps for seasonal malaria chemoprevention (SMC). They discuss how SMC campaigns work, the impact of USAID funding pauses on SMC campaigns, and GiveWell’s response to keep SMC campaigns on track.   Listen to Episode 2: Addressing Urgent Needs in Seasonal Malaria Chemoprevention   This situation is changing daily, and we’re constantly learning more. You can listen or subscribe to our podcast for the latest updates and read a summary of key takeaways from each podcast conversation on our blog. GiveWell has so far directed approximately