“Do you know what the most popular book is? No, it’s not Harry Potter. But it does talk about spells. It’s the Bible, and it has been for centuries. In the past 50 years alone, the Bible has sold over 3.9 billion copies. And the second best-selling book? The Quran, at 800 million copies.
As Oxford Professor William MacAskill, author of the new book “What We Owe The Future”—a tome on effective altruism and “longtermism”—explains, excerpts from these millennia-old schools of thought influence politics around the world: “The Babylonian Talmud, for example, compiled over a millennium ago, states that ‘the embryo is considered to be mere water until the fortieth day’—and today Jews tend to have much more liberal attitudes towards stem cell research than Catholics, who object to this use of embryos because they believe life begins at conception. Similarly, centuries-old dietary restrictions are still widely followed, as evidenced by India’s unusually high rate of vegetarianism, a $20 billion kosher food market, and many Muslims’ abstinence from alcohol.”
The reason for this is simple: once rooted, value systems tend to persist for an extremely long time. And when it comes to factory farming, there’s reason to believe we may be at an inflection point.”
Read the rest on Forbes.
Thanks for sharing Brian!
If you don't mind, I'll copy the two parts that stood out to me the most and helped to clarify the point for me better. If these points are valid, and I do think the logic makes sense, then this is quite concerning. Would love to hear other peoples thoughts on this.
I agree that this is an important issue and it feels like the time is ticking down on our window of opportunity to address it. I can imagine some scenarios in which this value lock in can play out.
At some point, AGI programmers will reach the point where they have the opportunity to train AGI to recognize suffering vs happiness as a strategy to optimize it to do the most good. Will those programmers think to include non-human species? I could see a scenario where programmers with human-centric world views would only think to include datasets with pictures and videos of human happiness and suffering. But if the programmers value animal sentience as well, then they could include datasets of different types of animals as well!
Ideally the AGI could identify some happiness/suffering markers that could apply to most nonhuman and human animals (vocalizations, changes in movement patterns, or changes in body temperature), but if they can’t then we may need to segment out different classes of animals for individual analysis. Like how would AGI reliably figure out when a fish is suffering?
And on top of all this, they would need to program the AGI to consider the animals based on moral weights, which we are woefully unclear on right now.
There is just so much we don’t know about how to quantify animal suffering and happiness which would be relevant in programming AGI. It would be great to be able to identify these factors so we can eventually get that research into the hands of the AGI programmers who become responsible for AI take-off. Of course, all this research could be for negligible impact if the key AGI programmers do not think animal welfare is an important enough issue to take on.
Are there any AI alignment researchers currently working on the issue of including animals in the development of AI safety and aligned goals?
Agree with the sentiment, thanks for the reply!
Exactly what I was thinking too. Unfortunately I think AGI will (and likely already is) move at light speed compared to the inclusion of animal consideration in our moral circle (when has tech not greatly exceeded pace with social movements?). If there's going to be a lock in, I'm fairly confident it's going to be well before we'll be where we need to be with our relationship with animals— even if we abolish factory farming by then.
So where does that leave us? Infiltrate companies working on AGI? Bring them into our circles and engage in conversations? Entice programmers/researchers with restricted grants (to help shape those datasets)? Physically mail them a copy of Animal Liberation? Are we even ready to engage in a meaningful way?
There's just so many questions. Really thought-provoking stuff.
Would love to know this too! I'm fairly new to this world and still poking around and learning, if I dig anything up I'll edit this post.
I'm currently working in technical AI safety, and I have two main thoughts on this:
1) We currently don't have the ability to robustly imbue AI with ANY values, let alone values that include all animals. We need to get a lot farther with solving this technical problem (the alignment problem) before we can meaningfully take any actions which will improve the longterm future for animals.
2) The AI Safety community generally seems mostly on board with animal welfare, but it's not a significant priority at all, and I don't think they take seriously the idea that there are S-risks downstream of human values (e.g. locking in wild-animal suffering). I'm personally pretty worried about this, not because I have a strong take about the probability of S-risks like this, but because the general vibe is just so apathetic about this kind of thing that I don't trust them to notice and take action if it were a serious problem.
Thanks for your comment. Are there any actions the EA community can take to help the AI Safety community prioritize animal welfare and take more seriously the idea that there are S-risks downstream or human values?
Archive Link
Are there many parts of the constitution that 'we', meaning people in general, find 'distasteful'? My impression is that most of the constitution either has few critics, or, if it has many critics, also has many defenders, or the critics disagree about how to change it. If we were to write it from scratch today, we'd probably end up with something quite different in many respects, but that doesn't mean there are massive generally agreed problems with it.
A majority oppose the electoral college (which gives unequal weight to different people). https://www.pewresearch.org/fact-tank/2022/08/05/majority-of-americans-continue-to-favor-moving-away-from-electoral-college/
Yes, there are many things that majorities want that they do not get, especially if there are many people who oppose the change, and they care a lot about opposing the change. A 60:40 split is a long way away from being sufficiently universal 'distaste' that we should expect it to necessarily triumph. This was true when the constitution was first written and remains the case today, so is not a sign on increasing rigidity.
Why then do you think there are fewer amendments overtime?
I think there are two main reasons:
Interesting. Thanks for your comments.
In the meantime, I would treat the constitution component in the piece as a metaphor to illustrate the idea of lock-in for a general audience.
I’d certainly write the constitution differently (why doesn’t it mention welfare for insects, for example?), but I more take it to mean that numerous amendments were required to make it moral, and still many more are needed.
Because most people do not care about insect welfare. The issue is not 'rigidity'; no sane amendment process would lead to the constitution mentioning insect welfare.