"Part one of our challenge is to solve the technical alignment problem, and that’s what everybody focuses on, but part two is: to whose values do you align the system once you’re capable of doing that, and that may turn out to be an even harder problem", Sam Altman, OpenAI CEO (Link).
In this post, I argue that:
1. "To whose values do you align the system" is a critically neglected space I termed “Moral Alignment.” Only a few organizations work for non-humans in this field, with a total budget of 4-5 million USD (not accounting for academic work). The scale of this space couldn’t be any bigger - the intersection between the most revolutionary technology ever and all sentient beings. While tractability remains uncertain, there is some promising positive evidence (See “The Tractability Open Question” section).
2. Given the first point, our movement must attract more resources, talent, and funding to address it. The goal is to value align AI with caring about all sentient beings: humans, animals, and potential future digital minds. In other words, I argue we should invest much more in promoting a sentient-centric AI.
The problem
What is Moral Alignment?
AI alignment focuses on ensuring AI systems act according to human intentions, emphasizing controllability and corrigibility (adaptability to changing human preferences). However, traditional alignment often ignores the ethical implications for all sentient beings. Moral Alignment, as part of the broader AI alignment and AI safety spaces, is a field focused on the values we aim to instill in AI. I argue that our goal should be to ensure AI is a positive force for all sentient beings.
Currently, as far as I know, no overarching organization, terms, or community unifies Moral Alignment (MA) as a field with a clear umbrella identity. While specific groups focus individually on animals, humans, or digital minds, such as AI for Animals, which does excellent community-building work around AI and animal welfare while
Until I read this article, saw this post and read the comments on it, I kind of imagined that EA's were very similar to normal people, just a bit more altruistic and a bit more expansive and maybe a bit more thoughtful.
This post scares the hell out of me.
This article is one of the worst articles I've ever seen in the NY Times. It is utter bullshit, but coated in meaningless, sweet-sounding words.
This is an attack on everything that we believe in! What the hell will it take to make EA's angry if this nonsense, in probably the most famous newspaper in the world, does not?
Why do we just sit back and think "that's not a very fair analysis"?
Does nobody feel an urgent need to defend ourselves, to get on TV and radio and places other than the EA forum and explain to the world that this article totally misses the point of EA, totally mischaracterises what we're trying to achieve and why?
If someone wrote an article about a minority group and described them with a few nasty racist stereotypes, there would be massive protests, retractions, apologies and a real effort to ensure that people were well informed about the reality.
The word "minority" is important here. If EA were the dominate mode of donating to charity, as it should be, then sure, it would be fine for someone to write that there is also value in donating to small, local charities, to challenge the status quo.
But EA represents only a small minority of donors today, so it is totally inappropriate for a journalist to pick on it.
But what really makes my blood boil are those who were not mentioned or consulted by this sad excuse for a journalist. For example, the people who desperately need food or medicine to survive. The animals who suffer in factory farms. The people who will suffer the most from climate-change.
We need to call this out for the bullshit it is. EA's believe that, when you donate, you should think a bit more about the people and animals who desperately need your help, and about what they need and how to help them, and maybe think a little bit less about the warm fuzzy feeling you get helping someone who will thank you profusely in person.
I absolutely refuse to accept that there is something wrong with that, and I find it shocking and appalling that the NY Times would publish this article as probably the only significant article they have published about EA since the last negative articles they published during the SBF affair.
At the very minimum, they have a responsibility to get their facts straight. Just read the four paragraphs where she introduces effective altruism. For her it is not a ground-roots movement, it is all about billionaires and ultra-wealthy. This is just not true. But she doesn't even mention that 99.999% of EA's are not rich by American standards - it's just that, unlike most, we're aware of how rich we are by global standards.
I would really hope to see a strong rebuttal submitted by someone in the EA movement. I would write it myself (and I will), but I don't think an article by me will get published in the NY Times. But there are people in the EA movement who are not millionaires but who do have the name-recognition and credibility to be listened to. This absolutely needs to happen, and fast. Maybe we could turn this negative into a positive. But giving season is already in full swing, and the people and animals who desperately depend on effective giving cannot afford to lose any of the insufficient donations they already get, even if it does mean that the local dog-shelter gets painted in bright Christmassy colours.
For now I plan to share this on my own social media and use it as an excuse to talk about effective giving and, as a side note, to share an example of shoddy journalism.