Hide table of contents

In a recent post, Scott Garrabrant gave an application of geometric rationality to the problem of work-life balance. Here's the setup: part of you wants to try to make the world better (I'll be calling this your "altruistic part") and part of you wants to relax and play video games (your "relaxation part"). Geometric rationality suggests doing a Nash bargain between the altruistic part and the relaxation part, across possible worlds that you might have found yourself in. In worlds where you're in an unusually good position to make the world better, the bargain commits you to spend most of your time doing that (satisfying your altruistic part); in return, in worlds where you're not in a good position to make the world better, you spend most of your time playing video games (satisfying your relaxation part).

Scott built a toy mathematical model and tested it with a nice example. The example involved five possible worlds, all equally likely. The way the math worked out, if you ranked the five worlds from least conducive to your altruism to most conducive, the Nash bargain had you spend 0%, 50%, 67%, 75%, and 80% of your time on altruism, respectively, in those five worlds. A pretty nice, intuitively satisfying list of numbers.

 

I recommend reading Scott's post before reading this one. And while I really liked the post, his example was constructed so that the probability distribution over how much altruistic impact you could have was pretty close to a uniform distribution. In the example, the world that was most conducive to your altruistic impact only gave you 1.7x more impact than the median world. But I think impact is much more unevenly distributed than that across possible worlds.

My issue with Scott's post wasn't his model (I quite liked the model!); it was only his example. So I decided to take the model and solve for the optimal bargain in more generality. I wanted to see what the results would be -- whether they'd still be intuitively compelling.

 

The rest of this post will be a bunch of math, but here's the (imprecise) TLDR:

  • If you know that your ability to make the world better is somewhat below average (across all possible worlds you could have found yourself in), you can spend all your time playing video games. If you know it's about average, you should aim for a fairly even mix of video games and altruism. If you know it's substantially above average (>10x, say), then you should spend almost all your time on altruism.
    • I mean "average" in the precise sense of the arithmetic mean -- not the median! There are plausible views under which you are in a top 0.001% world by impact potential, but still have below-average impact potential, because almost all the impact potential is concentrated in the top 0.00000001% of worlds.
  • (Epistemic status: super uncertain.) In practice, you have a huge amount of uncertainty about the distribution of your impact potential across worlds (not to mention uncertainty about how much impact potential you have in this world). While building this uncertainty into the model is hard, my intuition for geometric rationality suggests that a good way to deal with the uncertainty is to consider the probability that you'd assign to the proposition "my impact potential in this world is above average" (where the probability is over your logical uncertainty about the nature of the multiverse), and spend that fraction of your time improving the world.

 

All right, time to get to the math!

The setup

First, let  be the fraction of you that wants to improve the world (Scott's example uses ).

Order all worlds based on how impact potential you have, from least to most. For , let  be the impact potential you have in the -th quantile world (so  is an increasing function). (In Scott's example,  is  on  on  on  on , and  on .)

Let  be fraction of time you'll be devoting toward altruism in the -th quantile world. This is the function that we will be optimizing. Specifically, we will be looking for the function  that maximizes

.

(This is just the generalization of Scott's equation to a continuous setting.)

...okay, whoa there, what's this thing with the  in the exponent? You can read Scott's post on the geometric integral if you'd like. But if you (like me) don't have experience with geometric integrals, fear not: it'll go away once we take the log of that expression (which we can do because the function  that maximizes that expression will also be the  that maximizes its logarithm).

So, taking the log: we are looking for the function  that maximizes

(V stands for "value".)

The next section will be technical. Feel free to skip to the "Drawing conclusions" section below -- you won't lose too much.

 

Solving for the optimal function 

The right way to optimize this is with Lagrange multipliers. Unfortunately, I don't know Lagrange multipliers, so I do these sorts of optimization problems informally.

Let  be the optimal choice of . Let's consider making a small change to the value of  near a particular value : in particular, for very small  and very small positive , we will consider increasing  by  for . This change will approximately:

  • increase  by , where , and
  • decrease  by ,

for a net change of . In order for  to be optimal, this change cannot be positive. Thus, one of two things must be true for all : either

  • ; or
  • , but , so decreasing  is not possible.

(Briefly, we can't have the analogous edge case with  because we have  in a denominator.)

Therefore, rearranging terms, we have

.

We find it convenient to define , so that

.

(An aside: the quantity  (which, recall, is defined as ) has a nice interpretation: it's how much you improve the world, on average across worlds, if you follow the optimal bargain . The quantity  also has a nice interpretation: it's how much your relaxation part would need to be satisfied, if it were satisfied as much as your altruistic part, in proportion to how much these two parts exist.)

 

Our solution for  shows that you should spend all your time playing video games if you find yourself in a world where , and should spend at least some of your time saving the world if in fact .

...okay, but what is ? While we found  in terms of , we've defined  in terms of . That means we have an equation to solve! In particular, we have

By moving the  out of the integrand and rearranging, we can rewrite this as

.       (Eq. 1)

This isn't a solution, since there's a  on the right side of the equation. And we won't solve for  exactly, but we will prove a nice lemma that bounds  on both sides in terms of  -- which, remember, measures your average impact potential across possible worlds.

 

Lemma: .

Proof: the right-hand inequality is straightforward. We have

.

To prove the left-hand inequality, for convenience we will define , and . Then (from Eq. 1) . We also have that , i.e. that the average value of  on  is less than , since  and  is increasing. Therefore, we have

.

This completes the proof.

 

Now, recall that . Let . By Lemma 1, for any particular -value , we have

.

This gives us some nice bounds to work with when interpreting .

 

Drawing conclusions

Recall that  is the fraction of you that's altruistically minded. In this section, I recommend having a particular value of  in mind. Scott's choice of  is a reasonable one.

The optimal function  is . If you skipped the previous section, you don't need to worry about what  is; all you need to know is that we've bounded it: for every -value , we have

.

If  is such that , the model tells us to always play video games. We are guaranteed that this is the case whenever

.

The quantity on the left has a really natural interpretation: it is the factor by which you have more impact potential in world  than you do in the average world. So if you're in a somewhat below-average impact world, you can safely spend all your time playing video games (or so the model says).

On the other hand, what if this factor is instead pretty large -- at least , let's say? Then our bound tells us that  -- that is, you should spend nearly all your time working to improve the world.

And if for some reason you think that the factor is pretty close to  -- let's say it's exactly  -- then the time you should spend improving the world is somewhere between  and . (If , this gives a lower bound of 50% and an upper bound of 67%.) So you'd probably spend some intermediate fraction of your time improving the world.

 

Analysis and takeaways

First: how satisfying is this solution?

There's a sense in which I find it satisfying, which is that it accords with my intuition of what should have happened in the model. It seems like a reasonable way for the bargain to go down.

Then again, a sense in which I find it dissatisfying is that if you think (as I do) that the distribution of impact potential across possible worlds has a huge spread, then the bargain has almost all copies of you either spending all their time on video games or spending nearly all their time on improving the world. Which isn't really the takeaway that Scott was going for in his original post.

 

Is there a way to salvage this -- to restore the intuition that in fact you should have a reasonable work-life balance? I think the answer is yes! That's because you're probably very uncertain about what world you live in: you don't have a great idea about what the shape of  is, nor what quantile world you're in (i.e. what -value your world corresponds to). You should probably assign a decent chance to being in an above-average world in terms of impact potential, and a decent chance to being in a below-average world in terms of impact potential.[1]

Faced with this uncertainty, what do you do? If I understand the spirit of geometric rationality correctly, an important lesson is that you can do a Nash bargain not just between you-in-possible-worlds, but between different possibilities with respect to your epistemic state. For example, if you think there's a 90% chance that the Riemann hypothesis is true, you can bargain between you-if-the-Riemann-hypothesis-is-true and you-if-the-Riemann-hypothesis-is-false, even if there's an objective fact of the matter to which one is the actual you.

You can try to build this uncertainty into the model, but I think this would be really hard.[2] On the other hand, here's a basic fact about Nash bargaining: suppose there are two options,  and , and you assign x credence that  has utility  and  has utility , and  credence that  has utility  and  has utility , then the Nash bargain between these two epistemic states says to do A with probability  and  with probability .

And if you squint, this is sort of like what we have going on. Suppose you think that you're either in a substantially above-average world for impact potential (with probability ), or a substantially below-average one (with probability ). In the first epistemic state, "spend almost all your time altruistically" (option ) is much better than "spend all your time on video games" (option ); the the second case, option  is much better than option . And so -- if I'm not losing too much accuracy with my hand-wavy comparisons -- the Nash bargain between these two epistemic states would have you spend  fraction of your time on video games.

 

So suppose you think there's a 25% chance you live in a substantially below-average world in terms of impact potential, a 70% chance you live in a substantially above-average world, and a 5% chance of a close-to-average world -- well then, perhaps you should spend about 70% of your time working to improve the world!

...well, if you buy the basic model, and if you buy the hand-waving about Nash bargaining over epistemic states that I just did (which I'm not sure I buy)

...but, putting all these caveats to the side, I find this conclusion pretty satisfying.

 

  1. ^

    That last bit -- that maybe you're in a below average impact world -- is perhaps contrarian, but I stand by it. You may well think that you are in the top one-billionth of possible worlds in terms of impact potential. But it could be that among those worlds that are even higher than this one in terms of your impact potential, 1% of them have  times more potential beings than ours, and that this completely dominates the average.

  2. ^

    You'd need to have the domain of optimization be the set of your epistemic states instead of the set of world states. But your "epistemic state" includes way more than just a guess about what  looks like across worlds (or even a probability distribution over what  looks like across worlds). You need to include -- in each individual epistemic state -- your guesses about (what your probability distribution over what  looks like across worlds) looks like in all the other possible epistemic states. And you'll need to include your guesses about what those look like in all the other possible epistemic states. And so on. There are probably nice simplifying assumptions you could make to make this optimization problem tractable; I haven't thought much about this.

Comments2


Sorted by Click to highlight new comments since:

Interesting to think about! 

But for this kind of bargain to work, wouldn't you need confidence that the you in other worlds would uphold their end of the bargain? 

E.g., if it looks like I'm in videogame-world, it's probably pretty easy to spend lots of time playing videogames. But can I be confident that my counterpart in altruism-world will actually allocate enough of their time towards altruism?

(Note I don't know anything about Nash bargains and only read the non-maths parts of this post, so let me know if this is a basic misunderstanding!)

Great question -- you absolutely need to take that into account! You can only bargain with people who you expect to uphold the bargain. This probably means that when you're bargaining, you should weight "you in other worlds" in proportion to how likely they are to uphold the bargain. This seems really hard to think about and probably ties in with a bunch of complicated questions around decision theory.

Curated and popular this week
 ·  · 13m read
 · 
Notes  The following text explores, in a speculative manner, the evolutionary question: Did high-intensity affective states, specifically Pain, emerge early in evolutionary history, or did they develop gradually over time? Note: We are not neuroscientists; our work draws on our evolutionary biology background and our efforts to develop welfare metrics that accurately reflect reality and effectively reduce suffering. We hope these ideas may interest researchers in neuroscience, comparative cognition, and animal welfare science. This discussion is part of a broader manuscript in progress, focusing on interspecific comparisons of affective capacities—a critical question for advancing animal welfare science and estimating the Welfare Footprint of animal-sourced products.     Key points  Ultimate question: Do primitive sentient organisms experience extreme pain intensities, or fine-grained pain intensity discrimination, or both? Scientific framing: Pain functions as a biological signalling system that guides behavior by encoding motivational importance. The evolution of Pain signalling —its intensity range and resolution (i.e., the granularity with which differences in Pain intensity can be perceived)— can be viewed as an optimization problem, where neural architectures must balance computational efficiency, survival-driven signal prioritization, and adaptive flexibility. Mathematical clarification: Resolution is a fundamental requirement for encoding and processing information. Pain varies not only in overall intensity but also in granularity—how finely intensity levels can be distinguished.  Hypothetical Evolutionary Pathways: by analysing affective intensity (low, high) and resolution (low, high) as independent dimensions, we describe four illustrative evolutionary scenarios that provide a structured framework to examine whether primitive sentient organisms can experience Pain of high intensity, nuanced affective intensities, both, or neither.     Introdu
 ·  · 7m read
 · 
Article 5 of the 1948 Universal Declaration of Human Rights states: "Obviously, no one shall be subjected to torture or to cruel, inhuman or degrading treatment or punishment." OK, it doesn’t actually start with "obviously," but I like to imagine the commissioners all murmuring to themselves “obviously” when this item was brought up. I’m not sure what the causal effect of Article 5 (or the 1984 UN Convention Against Torture) has been on reducing torture globally, though the physical integrity rights index (which “captures the extent to which people are free from government torture and political killings”) has increased from 0.48 in 1948 to 0.67 in 2024 (which is good). However, the index reached 0.67 already back in 2001, so at least according to this metric, we haven’t made much progress in the past 25 years. Reducing government torture and killings seems to be low in tractability. Despite many countries having a physical integrity rights index close to 1.0 (i.e., virtually no government torture or political killings), many of their citizens still experience torture-level pain on a regular basis. I’m talking about cluster headache, the “most painful condition known to mankind” according to Dr. Caroline Ran of the Centre for Cluster Headache, a newly-founded research group at the Karolinska Institutet in Sweden. Dr. Caroline Ran speaking at the 2025 Symposium on the recent advances in Cluster Headache research and medicine Yesterday I had the opportunity to join the first-ever international research symposium on cluster headache organized at the Nobel Forum of the Karolinska Institutet. It was a 1-day gathering of roughly 100 participants interested in advancing our understanding of the origins of and potential treatments for cluster headache. I'd like to share some impressions in this post. The most compelling evidence for Dr. Ran’s quote above comes from a 2020 survey of cluster headache patients by Burish et al., which asked patients to rate cluster headach
 ·  · 2m read
 · 
A while back (as I've just been reminded by a discussion on another thread), David Thorstad wrote a bunch of posts critiquing the idea that small reductions in extinction risk have very high value, because the expected number of people who will exist in the future is very high: https://reflectivealtruism.com/category/my-papers/mistakes-in-moral-mathematics/. The arguments are quite complicated, but the basic points are that the expected number of people in the future is much lower than longtermists estimate because: -Longtermists tend to neglect the fact that even if your intervention blocks one extinction risk, there are others it might fail to block; surviving for billions  (or more) of years likely  requires driving extinction risk very low for a long period of time, and if we are not likely to survive that long, even conditional on longtermist interventions against one extinction risk succeeding, the value of preventing extinction (conditional on more happy people being valuable) is much lower.  -Longtermists tend to assume that in the future population will be roughly as large as the available resources can support. But ever since the industrial revolution, as countries get richer, their fertility rate falls and falls until it is below replacement. So we can't just assume future population sizes will be near the limits of what the available resources will support. Thorstad goes on to argue that this weakens the case for longtermism generally, not just the value of extinction risk reductions, since the case for longtermism is that future expected population  is many times the current population, or at least could be given plausible levels of longtermist extinction risk reduction effort. He also notes that if he can find multiple common mistakes in longtermist estimates of expected future population, we should expect that those estimates might be off in other ways. (At this point I would note that they could also be missing factors that bias their estimates of