Hide table of contents

TL;DR: Two different approaches for classifying projects or causes as “longtermist” clash significantly, which leads to oversimplifications of “longtermism” and causes people to pursue less impactful projects. I list some tentative suggestions at the end.

Notes: I discuss whether or not different projects are "longtermist" a lot in this post, but I don't want to imply that I think there's a very strong benefit to making sure that a project fits neatly in this group. For instance, I’m not arguing that people should be (or are) more “hardcore” with respect to longtermism, or anything like that. And if a project looks promising under multiple plausible worldviews, that should probably be considered a benefit (related, and also related).

Finally, arguments in this earlier post probably still inform my thinking on all of this.[1]

What is a “longtermist project”?

  1. One way of classifying something as “longtermist” is based on the definition of the philosophy/belief set. Under this approach: 
    1. A given project is “longtermist” (or not) depending on whether its primary purpose is to have an impact via influencing the long-term future — or at least if it looks good from a longtermist worldview. Not necessarily depending on whether it’s in biosecurity, animal welfare, AI, or whatever else. [3] 
  2. Alternatively, we could say that there are clusters of work (or causes) that are called longtermist — for historical reasons and because people who take a “longtermist” perspective often endorse working in these areas.[4] In this view:
    1. Biosecurity, AI safety, reducing the risk of global conflict, etc. are “longtermist causes.”  Extinction risk reduction generally falls under this, but also anything that looks like “reduce the risks of a catastrophic event,” even if the people pursuing the project are just hoping to prevent catastrophes from harming beings who are alive today. 

I’m sure there are other approaches to this classification, but I’ll focus on these two. 

These two approaches don’t always agree on what is “longtermist”

Below is a rough Venn diagram of some types of work that are called “longtermist” in one or both of the approaches (note that I didn’t try to classify things very carefully, or try to be exhaustive). 

Venn diagram: Motivated by longtermism vs. "longtermist" cluster. Only the former includes worrying about the stagnation of future economic progress. Only the latter includes developing better vaccines faster. And "both" includes AI alignment to prevent existential risks from AI.

Many of the relevant distinctions aren’t binaries, but rather spectrums  — some things look more or less “longtermist” under one of the two approaches, as opposed to simply being clearly “longtermist” or not — so a better picture might look more like a 2-dimensional visualization:

The X axis is "more motivated by longtermism" and the Y axis is "deeper in the longtermist cluster." 

In this diagram, the most canonical “longtermist” projects are in the top-right corner; the examples I chose are in the “longtermist cluster” and are often directly motivated by longtermism. This includes AI alignment, prevention of bioengineered pandemics, etc. Developing better flu vaccines is often associated with longtermism through biosecurity — an ostensibly “longtermist” cause — while advocacy or research about the moral relevance of non-human beings can be deeply motivated by longtermism but often isn’t thought of as “longtermist” because it’s not clearly in the cluster. 

Note that some people believe that basically no projects outside the canonical longtermist ones actually look promising under a longtermist worldview (although they might agree that the "longtermist cluster" contains things that also aren't "longtermist" under this classification). I currently disagree, but can see this point of view.

The difference between the two approaches is growing, and people who primarily think about the “longtermist cluster” now identify “longtermism” more closely with the causes it is currently associated with (AI safety, biosecurity, general risk reduction)

I think the two approaches have been different for a while (e.g. a lot of work on biosecurity is not motivated by longtermist thinking). 

And the approaches have diverged more recently. Some of the causes previously dominated by “longtermists[5] have grown a lot — more people have started to believe these causes are pressing problems and have started working on them (particularly AI risk and pandemic preparedness). These areas remain coded as “longtermist,” but many of the people working in them today are not significantly motivated by the desire to improve the long-term future. Relatedly, some risks that many believed were “long-term future risks” are now more widely viewed as urgent or near-term, so more people have started working on them for different reasons (there’s a convergence going on, to some extent; people with different values now share more priorities in practice). (See for instance this piece by Scott Alexander.)

Chart of Venn diagrams (motivated by longtermism vs. longtermist cluster) before and after the cluster grew. 

Some resulting problems

We assume someone’s philosophy based on their “revealed cause prioritization” and vice versa — which causes further difficulties

Examples of this kind of thing going wrong (some of these are very similar to things I've seen or heard about):

  1. Pluto says that they agree with the classic longtermist argument, so the people talking to Pluto (and, say, recommending opportunities to Pluto) assume that Pluto wouldn’t work on non-human welfare and don’t recommend those opportunities to them
    1. Or even just assume that Pluto wants to work on AI safety
  2. Ariel is working on AI safety, so people assume that she fully endorses longtermism or makes all cause-prioritization decisions based on longtermism (slightly related[6])
    1. ...which could additionally cause information cascades or incorrect deferral to Ariel in a way that causes more people to take on a kind of naive and more “hardcore” longtermist perspective on the world
  3. Eve doesn't buy the longtermist argument and gets frustrated because she sees prominent members of the community endorsing projects that are in the longtermist cluster a lot, which leads her to conclude that these people believe that longtermism is the only viable approach
  4. You see that Gascon is working on AI safety, so you assume that he cares about the long-term future and entrust him with something like a grantmaking project aims to improve the long-term future — but actually Gascon just finds his current AI safety project interesting, or thinks that AI is a near-term harm
  5. You meet Baymax, who works on biosecurity (say, far-UVC light), and you disagree with the longtermist argument. You start talking about cause prioritization and you feel like you can try to convince Baymax of all your reasons for not believing the longtermist case. This is not useful; Baymax's main crux for prioritizing biosecurity is the potential for near-term pandemic harm. 

People motivated by longtermism might overlook unusual areas of work or go for something that’s in the “longtermist cluster” but isn’t actually that promising when considered from a philosophically longtermist perspective

  1. When someone is seriously compelled by the longtermist argument, they might simply not consider areas like animal welfare, because they don’t see many discussions of animal welfare and longtermism; most “longtermist” spaces are full of discussions from the “longtermist cluster.” 
  2. If someone gets classified as “longtermist” or agrees with the classic argument and wants to do “longtermist work,” they might default to things in the “longtermist cluster” without prioritizing within that cluster enough. So they might e.g. do work that’s much more relevant for near-term epidemic preparedness than work that really helps in cases like global pandemics or non-bio work. 
    1. I have similar concerns about “EA projects” sometimes. 

I think some projects are promising under non-”longtermist” worldviews, but get somewhat overlooked by people who don’t think longtermism is sound as a philosophy. For example: 

  • Near-term AI safety work
  • Studying ways to improve international coordination to reduce the chances of war or conflict (because that would be bad for people alive today)
  • Developing vaccines for potential pandemics faster, developing cheaper and more accessible PPE, etc.
  • Reducing extinction risk because extinction is bad under completely different worldviews

I’m not sure how much this happens, in practice.

A few (tentative) suggestions/takeaways

I'm not sure about these! 

  1. Avoid implying that some project is only justified based on a pretty narrow philosophy/worldview (e.g. longtermism) unless you’re pretty sure that’s the case.
  2. I think generally “longtermist or not” is a worldview distinction that’s over-used in EA; e.g. human-focused vs. not is probably at least as important, as are more empirical questions like “do you believe that very powerful AI systems are coming soon?” I expect that it’s often more useful to group things based on other questions/classifications (instead of whether or not they’re “longtermist”). 
  3. When you do talk about whether something is “longtermist” or not, specify what your approach is and make that clear. (E.g. this was a useful conversation.) 
    1. I generally try to reserve the label “longtermist” for projects that are in fact motivated by longtermism (although at this point I also think it makes sense to use it to identify a cluster of work in our broader network). 
  4. Notice if you’re making assumptions about someone’s worldview based on something like the cluster they seem to be in (and vice versa), and check whether you endorse having this as a very strong prior. 
  5. Consider more often avoiding the term “longtermism” if you’re not really sure what you mean by it and how others will understand it. 
  1. ^

    I initially wrote a version of the current post over a year ago. In it, I mostly argued for a specific classification system (the motivation-based system below). I never ended up posting it because I wasn't convinced of my arguments, but related topics have come up many times since then and I've decided to try sharing something. I'm still not totally convinced of the details. 

    Relatedly, I think the arguments here are not particularly new, but I don't remember seeing someone lay them out in one place.

  2. ^

    You can also try to define projects as “longtermist” depending on where their actual impact lies, as opposed to the motivations of the people pursuing the project. I think this is more difficult but will result in a similar approach (although different conclusions). (I haven't checked carefully.)

  3. ^

    These projects can also look good under other philosophical/moral approaches — I’m not claiming that longtermism has to be the only reason for pursuing some kind of work.

  1. ^

    The cluster is also probably associated with requiring a higher willingness to work on low-probability events, longer feedback loops, etc. — things not necessarily relevant to the philosophy directly.

  2. ^

    which are therefore in the “longtermist cluster” for historical reasons

  3. ^

    Holden Karnofsky: There’s a line that you’ve probably heard before that is something like, “Most of the people we can help are in future generations, and there are so many people in future generations that that kind of just ends the conversation about how to do the most good — that it’s clearly and astronomically the case that focusing on future generations dominates all ethical considerations, or at least dominates all considerations of how to do the most good with your philanthropy or your career.” I kind of think of this as philosophical longtermism or philosophy-first longtermism. It kind of feels like you’ve ended the argument after you’ve pointed to the number of people in future generations.

    And we can get to this a bit later in the interview — I don’t think it’s a garbage view; I give some credence to it, I take it somewhat seriously, and I think it’s underrated by the world as a whole. But I would say that I give a minority of my moral parliament thinking this way. I would say that more of me than not thinks that’s not really the right way to think about doing good; that’s not really the right way to think about ethics — and I don’t think we can trust these numbers enough to feel that it’s such a blowout.

    And the reason that I’m currently focused on what’s classically considered “longtermist causes,” especially AI, is that I believe the risks are imminent and real enough that, even with much less aggressive valuations of the future, they are competitive, or perhaps the best thing to work on.

Show all footnotes
Comments3


Sorted by Click to highlight new comments since:

I like this framing, and I found that is resonates well with a minor level of discomfort I've had about uses of the term "longtermism." I especially liked the examples you have with characters (Pluto, Eve, etc.), as I find a few examples for a concept/framework to be very helping for grasping the "shape" of the idea.

Separately, I really like the graphics that you made for this. I think that you designed them well, from the color choices to the spacing. Could you share what tool/program you used to create these graphics?

Thank you! I used Google Slides. :)

Wow! I've never seen such nice design from Google Slides. Great job.

Curated and popular this week
 ·  · 5m read
 · 
[Cross-posted from my Substack here] If you spend time with people trying to change the world, you’ll come to an interesting conundrum: Various advocacy groups reference previous successful social movements as to why their chosen strategy is the most important one. Yet, these groups often follow wildly different strategies from each other to achieve social change. So, which one of them is right? The answer is all of them and none of them. This is because many people use research and historical movements to justify their pre-existing beliefs about how social change happens. Simply, you can find a case study to fit most plausible theories of how social change happens. For example, the groups might say: * Repeated nonviolent disruption is the key to social change, citing the Freedom Riders from the civil rights Movement or Act Up! from the gay rights movement. * Technological progress is what drives improvements in the human condition if you consider the development of the contraceptive pill funded by Katharine McCormick. * Organising and base-building is how change happens, as inspired by Ella Baker, the NAACP or Cesar Chavez from the United Workers Movement. * Insider advocacy is the real secret of social movements – look no further than how influential the Leadership Conference on Civil Rights was in passing the Civil Rights Acts of 1960 & 1964. * Democratic participation is the backbone of social change – just look at how Ireland lifted a ban on abortion via a Citizen’s Assembly. * And so on… To paint this picture, we can see this in action below: Source: Just Stop Oil which focuses on…civil resistance and disruption Source: The Civic Power Fund which focuses on… local organising What do we take away from all this? In my mind, a few key things: 1. Many different approaches have worked in changing the world so we should be humble and not assume we are doing The Most Important Thing 2. The case studies we focus on are likely confirmation bias, where
calebp
 ·  · 2m read
 · 
I speak to many entrepreneurial people trying to do a large amount of good by starting a nonprofit organisation. I think this is often an error for four main reasons. 1. Scalability 2. Capital counterfactuals 3. Standards 4. Learning potential 5. Earning to give potential These arguments are most applicable to starting high-growth organisations, such as startups.[1] Scalability There is a lot of capital available for startups, and established mechanisms exist to continue raising funds if the ROI appears high. It seems extremely difficult to operate a nonprofit with a budget of more than $30M per year (e.g., with approximately 150 people), but this is not particularly unusual for for-profit organisations. Capital Counterfactuals I generally believe that value-aligned funders are spending their money reasonably well, while for-profit investors are spending theirs extremely poorly (on altruistic grounds). If you can redirect that funding towards high-altruism value work, you could potentially create a much larger delta between your use of funding and the counterfactual of someone else receiving those funds. You also won’t be reliant on constantly convincing donors to give you money, once you’re generating revenue. Standards Nonprofits have significantly weaker feedback mechanisms compared to for-profits. They are often difficult to evaluate and lack a natural kill function. Few people are going to complain that you provided bad service when it didn’t cost them anything. Most nonprofits are not very ambitious, despite having large moral ambitions. It’s challenging to find talented people willing to accept a substantial pay cut to work with you. For-profits are considerably more likely to create something that people actually want. Learning Potential Most people should be trying to put themselves in a better position to do useful work later on. People often report learning a great deal from working at high-growth companies, building interesting connection
 ·  · 1m read
 · 
Need help planning your career? Probably Good’s 1-1 advising service is back! After refining our approach and expanding our capacity, we’re excited to once again offer personal advising sessions to help people figure out how to build careers that are good for them and for the world. Our advising is open to people at all career stages who want to have a positive impact across a range of cause areas—whether you're early in your career, looking to make a transition, or facing uncertainty about your next steps. Some applicants come in with specific plans they want feedback on, while others are just beginning to explore what impactful careers could look like for them. Either way, we aim to provide useful guidance tailored to your situation. Learn more about our advising program and apply here. Also, if you know someone who might benefit from an advising call, we’d really appreciate you passing this along. Looking forward to hearing from those interested. Feel free to get in touch if you have any questions. Finally, we wanted to say a big thank you to 80,000 Hours for their help! The input that they gave us, both now and earlier in the process, was instrumental in shaping what our advising program will look like, and we really appreciate their support.