This seems like a not at all new or original idea, but maybe worth explicitly pointing out. 

A lot of times at EA events I've heard something along the lines of "as an EA, I'm trying to maximize the amount of good I accomplish."

But this isn't quite what we should (in theory, assuming I don't care about my non-EA goals, which I totally do) be doing. Instead, as an EA I'm trying to maximize how awesome the world over time will be, not just the "awesomeness" that can be attributed to me. 

More concretely, doing things like research, entrepreneurship, etc. are all shiny things that demonstrate that I've done lots of good. On the other hand, for AI Safety field building, spreading EA in areas where EA isn't really present, etc., the impact here can't always be traced back directly to me, but these also seem incredibly important and useful. I'm suprised at how many more safety researchers than field builders there are, and I'm guessing one reason is that it's difficult to pinpoint how much impact you've actually personally made as a field builder. 

Most roles would probably also look slightly different if people focused less on their personal contribution to the world. For example, it might make sense for a researcher to give a cool project idea to a not-currently-occupied person who they know will be better at the project.

 I think it's worth the effort to consciously try to feel tons of happiness from seeing altruism being accomplished, and not just from accomplishing altruism. 

42

0
0

Reactions

0
0
Comments7


Sorted by Click to highlight new comments since:

But ultimately, I think you are basically saying you are trying to maximise your counterfactual impact, no? Not the impact that can be traced to you, but the impact that you have, through all channels.

In the context of this post, I read "my contribution to good" to mean "good done that is clearly attributed to me" rather than "my counterfactual impact".

 Though I'd also usually think "my contribution"  is my "counterfactual impact", I still think this reframing ("I am maximizing how much good I do" to "I am maximizing how good the world is") might be instrumentally very useful for feeling good about more indirect ways of having a counterfactual impact.

yeah; it seems obvious to me that "the good I accomplish" includes my contribution to allowing others to do good. I'm open to seeing evidence but I suspect the reason field-building, movement-building etc. isn't done as much as OP would like has nothing to do with this kind of confusion. In fact I think it's questionable how much you can do at the meta level if your direct work doesn't measure up. People show up when they see cool stuff being done, not so much when they hear you talk about the cool stuff that someone else should do. Sputnik did a great deal more for science and engineering education than running a bunch of commercials about the importance of science would have. 

not so much when they hear you talk about the cool stuff that someone else should do.

Maybe that feels a bit unfair non-steelmanny to me? There are other ways of motivating and helping others and the process other than just saying 'wouldn't it be great if someone solved the alignment problem' etc.

Such as:

  • Encouraging people who are working on the problem
  • Providing inputs and support to others working on important problems
  • Offering career advice
  • Helping communicate and explain the work that is being done, in term helping people coordinate

Yeah, this makes sense. That being said, I'm guessing while some people in theory are trying to maximize the "good" they accomplish, in practice it's easy to forget about options that aren't easily traceable. My point was also that it's worth explicitly putting in effort  to look for these kinds of options. 

By options, I mean something like giving a research project to a more capable person. I'm guessing some people wouldn't consider that this is a thing they can do. 

I like that you shared this thought. It parallels with my understanding of servant leadership, and the humility of it appeals to me. I do think that all of us (myself included) can get easily caught up in the ego of accomplishing things, of wanting others to respect us, of wanting to be admired. But I agree with you: if our goal is to make the world a better place, it doesn't really matter if I make it better place or if I spend my time allowing/enabling/supporting someone else to make it a better place.

 

Instead of "I'm trying to maximize the amount of good I accomplish" I'd love to see more people adopt the mindset of "I'm trying to maximize the amount of good accomplished."

This is also captured by the 'hits-based' framing, where many people will try and fail, but some will have huge successes.

Curated and popular this week
 ·  · 25m read
 · 
Epistemic status: This post — the result of a loosely timeboxed ~2-day sprint[1] — is more like “research notes with rough takes” than “report with solid answers.” You should interpret the things we say as best guesses, and not give them much more weight than that. Summary There’s been some discussion of what “transformative AI may arrive soon” might mean for animal advocates. After a very shallow review, we’ve tentatively concluded that radical changes to the animal welfare (AW) field are not yet warranted. In particular: * Some ideas in this space seem fairly promising, but in the “maybe a researcher should look into this” stage, rather than “shovel-ready” * We’re skeptical of the case for most speculative “TAI<>AW” projects * We think the most common version of this argument underrates how radically weird post-“transformative”-AI worlds would be, and how much this harms our ability to predict the longer-run effects of interventions available to us today. Without specific reasons to believe that an intervention is especially robust,[2] we think it’s best to discount its expected value to ~zero. Here’s a brief overview of our (tentative!) actionable takes on this question[3]: ✅ Some things we recommend❌ Some things we don’t recommend * Dedicating some amount of (ongoing) attention to the possibility of “AW lock ins”[4]  * Pursuing other exploratory research on what transformative AI might mean for animals & how to help (we’re unconvinced by most existing proposals, but many of these ideas have received <1 month of research effort from everyone in the space combined — it would be unsurprising if even just a few months of effort turned up better ideas) * Investing in highly “flexible” capacity for advancing animal interests in AI-transformed worlds * Trying to use AI for near-term animal welfare work, and fundraising from donors who have invested in AI * Heavily discounting “normal” interventions that take 10+ years to help animals * “Rowing” on na
 ·  · 3m read
 · 
About the program Hi! We’re Chana and Aric, from the new 80,000 Hours video program. For over a decade, 80,000 Hours has been talking about the world’s most pressing problems in newsletters, articles and many extremely lengthy podcasts. But today’s world calls for video, so we’ve started a video program[1], and we’re so excited to tell you about it! 80,000 Hours is launching AI in Context, a new YouTube channel hosted by Aric Floyd. Together with associated Instagram and TikTok accounts, the channel will aim to inform, entertain, and energize with a mix of long and shortform videos about the risks of transformative AI, and what people can do about them. [Chana has also been experimenting with making shortform videos, which you can check out here; we’re still deciding on what form her content creation will take] We hope to bring our own personalities and perspectives on these issues, alongside humor, earnestness, and nuance. We want to help people make sense of the world we're in and think about what role they might play in the upcoming years of potentially rapid change. Our first long-form video For our first long-form video, we decided to explore AI Futures Project’s AI 2027 scenario (which has been widely discussed on the Forum). It combines quantitative forecasting and storytelling to depict a possible future that might include human extinction, or in a better outcome, “merely” an unprecedented concentration of power. Why? We wanted to start our new channel with a compelling story that viewers can sink their teeth into, and that a wide audience would have reason to watch, even if they don’t yet know who we are or trust our viewpoints yet. (We think a video about “Why AI might pose an existential risk”, for example, might depend more on pre-existing trust to succeed.) We also saw this as an opportunity to tell the world about the ideas and people that have for years been anticipating the progress and dangers of AI (that’s many of you!), and invite the br
 ·  · 14m read
 · 
Cross-posted from my personal blog after strong encouragement from two people. The text is partially personal musings over my journey through the pre-surgery process, and partially a review of arguments why it might or might not make sense for a person who cares about effective altruism to donate a kidney. Some of the content is specific to where I am based in (Finland), and some of the content is less honed than I'd like, but if I didn't timebox my writing and just push whatever I have at the end of it I probably would not publish anything. Hope you enjoy! June 4, 2025 I was eating phở in the city center when the call came. The EM study had been done. "Bad news," my nephrologist said. "You have a kidney disease." The words hit harder than the spicy broth. After more than nine months of tests, blood draws, and even a kidney biopsy, my journey to donate a kidney to my friend had just come to an abrupt end. Let’s rewind to where this all began. The Decision When Effective Altruism volunteer Mikko mentioned that his kidney transplant was showing signs of chronic rejection, I jumped and offered him my kidney. I had read Dylan Matthew’s kidney donation story, Scott Alexander’s kidney donation story, and had taken part in a discussion on estimating the level of kidney demand in Finland in a local EA discussion group that had at least one Finnish doctor involved. The statistics were reassuring. Kidney donations are very safe with a perioperative death rate of around 0.03% to 0.06%. You will have a slightly increased risk of kidney-related issues later in life (1-2% likelihood of kidney failure), and there is also some weak evidence that suggests people who donate a kidney have very slightly shorter life expectancy than those who have both kidneys intact. The QALY Trade-off There are multiple attempts to calculate what the quality-adjusted life year (QALY) trade for a kidney donation actually is, and to me it looks like donating a kidney is by default similar to