Take the 2025 EA Forum Survey to help inform our strategy and prioritiesTake the survey
Hide table of contents

Epistemic status: uncertain about whether this accurately describes Ben's views. The podcast is great and he's also doing a very interesting AMA. This is a very complex topic, and I would love to hear lots of different perspectives and have them really fleshed out in detail. The below is my attempt at a quick summary for those short on time.

Core ideas I took away (not that I necessarily all agree with)

  1. Brain in a box - the classic Bostrom-Yudkowsky scenario is where there’s a superintelligent AGI developed which is far more capable than anything else people are dealing with, i.e. a brain in a box, but actually we’d expect systems to develop incrementally and so we should have other examples of similar concrete problems to work on.
  2. Intelligence explosion - one of the concepts behind the runaway intelligence explosion is that a system is recursively self-improving, so the AI starts to rewrite its own code or hardware and then get much better. But there are many tasks that go into system improvement, and even coding requires many different skills, so just because a system might be able to improve one of its inputs, that doesn’t mean that its overall capacity should increase.
  3. Entanglement and capabilities - when we’ve had AI systems they’ve usually got more capable by getting better at giving us what we want, so by exploring the potential space of solutions more and more carefully. For example house cleaning robots only get better as they learn more about our preferences. Thermostats only get more effective and capable when they get better at moderating the temperature, because the intelligence of meeting the goal is entangled with the goal itself. This should make us suspicious of extremely powerful and capable systems that also have divergent goals to ours.
  4. Hard to shape the future - if we take these arguments seriously, it also might be the case that AI safety can develop more gradually as a field over coming decades, and that while it’s important, it just might not be as much of a race as some have previously argued. To take something potentially analogous, it’s not clear what someone in the 1500s could have done to influence the industrial revolution, even if they had strong reasons to think it would take off.

Some other points

  • In the interview, Ben also mentions that there are “multiple salient emerging forms of military technology” which could be of similar importance, giving the example of hypersonic glide vehicles. I’ve considered taking this course in Science and International Security at the War Studies department at KCL, and I’ve uploaded the syllabi for the main units here and here. Other examples are space security and cyber security.
  • In the 80K podcast with Stuart Russell, Stuart calls out Rob for conflating ML systems with AI systems. Just to define terms, machine learning systems improve automatically through experience with data. And artificial intelligence is a much broader area of research, including robotics, computer vision, classical search, logical reasoning, and many other areas. Stuart makes the point in the podcast that Google’s self-driving cars mostly use classical search, and so only looking at ML is part of the picture.
  • Rohin Shah reviews Ben’s interview favourably in the Alignment Newsletter here.
  • Rohin also discusses AI safety with Buck Shlegeris here, but I haven’t finished the interview (though I did find the discussion quite confrontational and switched off)
  • I’ve also just pulled out the most contentious points - Ben gives a very rounded and considerate interview, which I’d recommend listening to in full

My takeaways

  • I found Ben’s arguments to be very useful and interesting
  • I agree that working on existential risk involves more than just one technology, and so there could be fruitful work in security studies and power structures, with a great popular example of theoretical work being Destined for War (also see The Vulnerable World Hypothesis). This work seems important and neglected.
  • While I think Ben’s arguments require responses from people looking into AI, from my perspective the main idea that humans are not optimally intelligent, and that more advanced technologies could exploit that significantly in the future to produce undesirable outcomes (including human extinction and s-risks) seem plausible to me
  • Before listening to this podcast, I’d have put a 10-30% chance on a <6 month hard take-off scenario this century, conditional on AI safety work happening and the world not being radically different from now, but I’d now put it at something like 5-20%, though I’m really not an expert here, so would expect my views to change a lot (immediately after reading Superintelligence I was probably at 50%).
  • I was also very glad that this perspective was aired, and I hope it leads to more fruitful discussions
  • At the end of his slides, Ben closes with an important point for the EA community, ‘If we’ve failed to notice important issues with classic arguments until recently, we should also worry about our ability to assess new arguments.’

10

0
0

Reactions

0
0

More posts like this

Comments1


Sorted by Click to highlight new comments since:

"To take something potentially analogous, it’s not clear what someone in the 1500s could have done to influence the industrial revolution, even if they had strong reasons to think it would take off."

Detail - Ben says:

At the same time though, if you’re living in 1750 or something and you’re trying to think, “How do I make the industrial revolution go well? How do I make the world better… let’s say in the year 2000, or even after that”, knowing that industrialization is going to be very important… it’s not really clear what you do to make things go different in a foreseeably positive way.

So he's talking about the start of the Industrial Revolution, not two centuries before.

More from Ben
Curated and popular this week
 ·  · 1m read
 · 
This morning I was looking into Switzerland's new animal welfare labelling law. I was going through the list of abuses that are now required to be documented on labels, and one of them made me do a double-take: "Frogs: Leg removal without anaesthesia."  This confused me. Why are we talking about anaesthesia? Shouldn't the frogs be dead before having their legs removed? It turns out the answer is no; standard industry practice is to cut their legs off while they are fully conscious. They remain alive and responsive for up to 15 minutes afterward. As far as I can tell, there are zero welfare regulations in any major producing country. The scientific evidence for frog sentience is robust - they have nociceptors, opioid receptors, demonstrate pain avoidance learning, and show cognitive abilities including spatial mapping and rule-based learning.  It's hard to find data on the scale of this issue, but estimates put the order of magnitude at billions of frogs annually. I could not find any organisations working directly on frog welfare interventions.  Here are the organizations I found that come closest: * Animal Welfare Institute has documented the issue and published reports, but their focus appears more on the ecological impact and population decline rather than welfare reforms * PETA has conducted investigations and released footage, but their approach is typically to advocate for complete elimination of the practice rather than welfare improvements * Pro Wildlife, Defenders of Wildlife focus on conservation and sustainability rather than welfare standards This issue seems tractable. There is scientific research on humane euthanasia methods for amphibians, but this research is primarily for laboratory settings rather than commercial operations. The EU imports the majority of traded frog legs through just a few countries such as Indonesia and Vietnam, creating clear policy leverage points. A major retailer (Carrefour) just stopped selling frog legs after welfar
 ·  · 1m read
 · 
The Open Wing Alliance 2025 cage-free fulfillment report shows that 92% of cage-free egg commitments with deadlines of 2024 or earlier have been fulfilled, up from 89% last year. The report details a regional breakdown of progress around the world.  Thanks to OWA groups and others for continuing to hold companies around the world accountable for their commitments.
 ·  · 4m read
 · 
Note: This post was crossposted from the Open Philanthropy Farm Animal Welfare Research Newsletter by the Forum team, with the author's permission. The author may not see or respond to comments on this post. ---------------------------------------- > Why ending the worst abuses of factory farming is an issue ripe for moral reform I recently joined Dwarkesh Patel’s podcast to discuss factory farming. I hope you’ll give it a listen — and consider supporting his fundraiser for FarmKind’s Impact Fund. (Dwarkesh is matching all donations up to $250K; use the code “dwarkesh”.) We discuss two contradictory views about factory farming that produce the same conclusion: that its end is either inevitable or impossible. Some techno-optimists assume factory farming will vanish in the wake of AGI. Some pessimists see reforming it as a hopeless cause. Both camps arrive at the same conclusion: fatalism. If factory farming is destined to end, or persist, then what’s the point in fighting it? I think both views are wrong. In fact, I think factory farming sits in the ideal position for moral reform. Because its end is neither inevitable nor impossible, it offers a unique opportunity for advocacy to change the trajectory of human moral progress. Not inevitable Dwarkesh raised an objection to working on factory farming that I often hear from techno-optimists who care about the issue: isn’t its end inevitable? Some cite the long arc of moral progress; others the promise of vast technological change like cultivated meat or Artificial General Intelligence (AGI) which surpasses human capabilities. It’s true that humanity has achieved incredible moral progress for humans. But that progress was never inevitable — it was the result of moral and political reform as much as technology. And that moral progress mostly hasn’t yet extended to animals. For them, the long moral arc of history has so far only bent downward. Technology may one day end factory farming, just as cars liberated w