This week, we are highlighting Forethought's Better Futures series. To make the future go better, we can either work to avoid near-term catastrophes like human extinction or improve the futures where we survive. This series from Forethought explores that second option.
Fin Moorhouse (@finm), who authored two chapters in the series (Convergence and Compromise, and No Easy Eutopia) along with @William_MacAskill, has agreed to answer a few of your questions.
You can read (and comment) on the full series on the Forum. In order, the chapters are:
- Introducing Better Futures
- No Easy Eutopia
- Convergence and Compromise
- Persistent Path-Dependence
- How to Make the Future Better
- Supplement: The Basic Case for Better futures
Leave your questions and comments below. Note that Fin isn't committing to answer every question, and if you see someone else's question you can answer, you're free to.

Given this combination of views, I'm surprised that Will doesn't support what @Holly Elmore ⏸️ 🔸 calls "Pause NOW" and instead want to see a pause later (after we have human-level AI). I'm curious if your own views are similar or how they differ from Will's. (My own "expected value of the future, given survival" I would say is similarly pessimistic, but I'm reluctant to put into numbers due to being very unsure how to quantify it.)
Aside from what Holly said in the linked comment, which I agree with, another argument more relevant to the current discussion is that many opportunities for making the future better seem to exist during the AI transition, including the early parts of it, so by not pausing ASAP (and currently having few resources for such interventions), we're permanently giving up these opportunities. Conversely, by pausing NOW, we buy more time to think and strategize about how to better intervene on these opportunities, or otherwise lay the groundwork for them.
For example, during the pause, we could:
Such interventions could mean the difference between the first human-level AIs being competent and critical moral/philosophical advisors, or independent moral (and safe) agents, vs uncritically doing what humans seem to want and/or giving bad/incompetent/sycophantic "advice" (when humans think to ask for it), which seemingly can make a big difference to how well the future goes.
What do you think about this argument, and overall about pause now vs later?
Hi Fin,
I have a lot of questions so I figure I would just share all of them and you could respond to the ones you want to.
Thanks James!
Some things I appreciate in my colleagues: having some discernment for which questions or ideas are most important, rather than just conceptually interesting but not urgent; being able to contribute to group conversations by driving at cruxes, being willing to ask naive questions and avoiding the impulse to sound clever for the sake of it, and being able to spot and entertain "big if true" hypotheses; and being able to clearly communicate ideas where you often don't have an especially deep literature to draw on.
I think it's important to understand how AI works in the fundamentals; including some of the theory. I don't think it's important to have deep technical knowledge of LLMs, unless you can think why those details could end up being relevant for macrostrategy.
On your second question, many of those points seem good to me. I'll single out "Locking-in one’s values" since I've been thinking about it recently. It seems to me that some people roughly think that great futures are futures which resemble our own (or which carry on our values) in many particular ways. In particular, maybe great futures are futures which are recognisably human in their values. Inhuman futures, like futures where AI successors call the shots, might just seem empty of what we today care about; even if they involve a lot of moral reflection and nothing morally offensive from a human perspective. We could call this a "humanity forever" view.
On the other hand, some people roughly think that great futures are necessarily futures which are radically different from humanity today, including in the values which guide it, and perhaps the kind of actors living there. See Dan Faggella on the "Worthy Successor" idea (and here), which I see as one version of this view.
Both these views care about preventing obvious catastrophes from AGI, but it seems to me like they might end up disagreeing quite profoundly on what should come next. It's possible that there is opportunity for trade and compromise between the two views, but in any case this strikes me as a potentially important difference in approach to post-AGI futures.
Firstly, you're right that the series doesn't discuss negative futures, but I should say that's not because Will or I think they are worth ignoring, or very unlikely in absolute terms. We didn't discuss them more just so we could make a more focused argument about how to think about making good futures even better.
I think your point (quoted) touches on the difference I mentioned above between "humanity forever" views and views which are more open to change in values. I think it's coherent to take a view such as:
Better Futures argues that these views may be less tenable than they first appear, but I think they're not totally doomed.
Additionally I would point out a potential "missing mood" in the framing we adopt of cardinally quantifying the value of the future in terms of a fraction of the value of the best feasible future. This suggests futures which are only, say, 10E-5 the value of the best feasible future are barren, hollow, 'neutral'. But this would be a mistake: potentially our own world, even with all the harm and pain removed, is achieving a tiny fraction of what a great future could achieve. So we might imagine (as Better Futures points out) a "common-sense eutopia" which is radically better than the world today, but still only a fraction as good as things could get. That could be true, but it doesn't undermine the value of such a future, which would also truly be (by stipulation) wildly better than the world today! All the joy and freedom and discovery and so on, in this near-zero world, would be entirely real and could dwarf all the good we have achieved and enjoyed so far.
Maybe I'm misreading but I don't think it follows from uncertainty about how things go that many different things will actually happen. For example, if you're uncertain who wins a political election, you don't infer that everyone wins and shares power.
I'm in a few minds about this, so I'll just list some reactions:
We are at Forethought, running a research programme on space right now, which I guess reflects a view that it does seem worth investigating more. I don't think the central case for space runs through the hope for binding international treaties because I agree that we shouldn't expect them to hold. I think there are a few other reasons to want to investigate space. One is that the space economy could be somewhat relevant for the course of AGI development, for example if orbital data centres are a big deal, or because of the role of sensing satellites in peace and security.
Another is that most of the physical stuff is in space. At some point it seems likely to me, if the human project continues at all, that most of the important stuff will also eventually be in space. AGI + automated manufacturing + rapid R&D progress suggests that expanding into space could happen in the time span of decades rather than centuries or millennia; and that seems generically worth planning for. And it seems like there are some policy levers which don't root through international treaties.
To be clear I don't currently think that space governance should be the next big cause in EA or anything like that.
This feels like a slightly odd sentence construction, because you seem to be saying that longtermism is unhelpful because it requires people to believe one of its central claims. I agree it's contentious and I'm certainly not confident that the effects of our actions could persist for millions of years but it seems plausible enough that the anticipated long-term effects of our actions should meaningfully weigh into what we prioritise, at least where you can tell a story about how your decisions could have some systematic long-run effects.
I do think that is plausible. Although, to state the obvious, there is a difference between which ideas have good or bad PR effects when you say them out loud, and which ideas are actually true or important. So questions about communicating longtermist ideas are, naturally, different from the question of whether longtermist ideas are worth taking seriously as ideas.
And then, I also want to say: the full-on version of longtermism — that the very long-run effects of our actions are overwhelmingly important for what we prioritise — just doesn't feel especially necessary for working on most or even all of the topics that Forethought is focused on. There is a far more common-sense and mundane reason to focus on them, which is that they could matter enormously within our own lifetimes! Another way of putting that is that when trying to prioritise between possible focuses within Forethought, my personal view is that longtermism is rarely a crux. Maybe my colleagues disagree with that; obviously I'm not speaking on their behalf.
I'm not sure I'm entirely following your points but I don't see a strong reason why AIs or non-human entities could not in principle engage in genuine moral reasoning in the same way that humans do. Maybe instead the AIs will do something which superficially resembles real moral reasoning, but which is closer to just telling humans what they want to hear.
I do think that is not a crazy thing to worry about because it is much easier to train some skill where an uncontroversial and abundant source of ground truth data exists. Moral reasoning is not one of those domains because people often don't agree on what good moral reasoning looks like. So I think there is much work to be done on that front although I'm not sure that answers your question.
Thanks again for your questions!
Hey Fin,
Thanks for so thoughtfully answering my questions!
Forethought's view that improving the future conditional on survival is more important than ensuring survival goes against the dominant view in EA for many years that we need to reduce extinction risk. Two questions on this:
I really liked the series :)