Hide table of contents

Epoch researchers recently shared a great blog post titled “Most AI value will come from broad automation, not from R&D”. You should read it! This seems like a big-if-true claim given how much focus goes into the ‘intelligence explosion followed by an industrial explosion’ narrative. Below, I summarise their post and then offer some critiques and responses.

Summary of Epoch

In the post, Ege Erdil and @Matthew_Barnett  argue (contrary to the standard narrative) that automating R&D tasks, in particular AI R&D, will not be central to the economic significance of AI. Instead, they argue general diffusion of AI through the economy by automating non-research jobs will be the more important source of economic growth. This is because:

  • Historical evidence. R&D has only contributed about 20% of labour productivity growth in the US since 1988. Some of the main non-R&D factors are capital accumulation (buying more and fancier machines can make the same worker more productive) and knowledge diffusion (e.g. learning new techniques and institutional practices from more productive firms).
    • This gives us a (maybe weak) prior that R&D isn’t the main deal. Indeed, the US spends about $5T on capital accumulation and $1T on R&D, roughly proportional to the estimated shares of productivity gains.
  • Software singularity may fizzle. A common reason to think that AI will be different than the historical trend of R&D playing a small role in productivity growth is that we will automate AI R&D itself, get recursive improvements to reach superintelligence, and only then rapidly automate all other remote work. Erdil and Barnett think this is possible, but somewhat unlikely, given that algorithmic insights may get progressively harder to find at a steep rate.
    • In this scenario, compute and data are not growing very rapidly, so progress must come from algorithms. Probably, there will be strong complementarity between data/compute and algorithms, such that just better algorithms can only very imperfectly substitute for more data and compute.
    • Historically, automating parts of R&D (e.g. calculations and, to some extent, coding) has not led to dramatic scientific acceleration.
  • R&D is hard. Non-AI R&D often requires significant physical labour (e.g. in wet labs for biology) which will be slow to automate. Even AI R&D requires agentic computer-use skills (and long-term planning and research taste) which will be slower to automate than pure abstract reasoning tasks. As such, by the time AIs are able to automate much of R&D, they will have already been previously able to automate many other remote jobs, and will be broadly diffused in the economy. This broad diffusion will be the main driver of economic growth.

A key strategic implication of this is that we will likely have widespread AI diffusion contributing a significant fraction of GDP before we have recursive improvement and superintelligence.

Response to Epoch

Epistemic note: I have thought about this probably a lot less than Epoch people, and also have a less relevant background. Which makes me think I should defer at least moderately to them.

But it still seems valuable to give my object-level thoughts. I mainly don’t directly dispute Epoch’s points, instead giving countervailing reasons that push against their conclusion.

Firstly, I think AI R&D may be relatively early to be automated.[1] It seems very likely that many physical jobs (e.g. nursing) will be automated relatively late, but even compared to other fully remote jobs:

  • AI researchers command high salaries. Particularly if inference-compute scaling continues, the first automated remote workers may be quite expensive to run.[2] As such, supply may be quite limited, and the first use cases would (all else equal) be in fields with very high human salaries. AI researchers often have very high salaries, so replacing them will be a top priority.[3]

  • R&D skill scales strongly with cognitive ability.[4] We may get models that are weakly superhuman abstract reasoners before we have models that are good at doing a wider range of computer and physical tasks. R&D jobs seem to have some of the best returns to intelligence - e.g. a 2 standard deviations smarter researcher might be vastly more innovative and productive than her average colleague. This seems different to many other professions where raw intelligence seems like a smaller fraction of the overall job.[5]

  • AI R&D is amenable to trial and error. We may get models that have flashes of brilliance before we get models that reliably never make mistakes. Science is a strong-link problem where what matters most is your very best ideas, not how many mistakes you make or bad ideas you come up with.[6] Fields like surgery that require very high robustness may be automated later than R&D.

  • AI R&D is not heavily regulated. Plenty of fields already do or likely will require humans to do particular tasks. Since AI R&D is less regulated, that will make it easier to automate early.
  • AI companies are experts in AI R&D. It is easier to automate what you know lots about, and have proprietary internal training data (e.g. history of GitHub repos). As Erdil and Barnett note, leading AI companies seem bullish on automating AI R&D. Maybe this is misguided or just hype, but I place a bit more weight on that being an honest assessment of AI companies' plans. These predictions may also be self-fulfilling, as AI companies could have quite a bit of discretion about where to direct their automation efforts first.

Secondly, even if broad diffusion of AI through the economy contributes more to general economic growth, I think R&D might be more important strategically. In particular, even if broad automation of labour throughout the economy leads to higher economic growth, tightly scoped automation of R&D in strategic sectors - AI, chip design, cyber offense/defense, military technology - would lead to a larger increase in national power. Competitive geopolitical pressures will likely force countries to focus first on bolstering their military-industrial might rather than broader consumer welfare.[7] 

Thirdly, the post critiques a ‘software intelligence explosion’ but does not discuss a ‘chip technology’ IE. As Davidson, Hadshar, and MacAskill argue, automating chip technology research is another feedback loop that could, together with software, create accelerating progress. The third type of ‘full stack’ IE involving general capital accumulation and investment in semiconductor manufacturing is closer to what Erdil and Barnett think is likely.

Overall, I still think it is more likely than not that targeted AI R&D automation will be the main game, and broader labour automation will be downstream of this and less strategically important.

Thanks to Nathan Barnard, Ryan Greenblatt, and Rose Hadshar for helpful comments.

  1. ^

     Tom Davidson has raised some similar points here.

  2. ^

     However, inference costs tend to decline rapidly over time.

  3. ^

     Other fields with huge salaries, like quant trading, will also presumably be focuses of early remote work automation. This would fit in with the trend of increasingly automated trading anyway.

  4. ^

     @Jackson Wagner  made some similar useful points in a comment on the blog post.

  5. ^

     E.g. consider an architect. This job seems liable to be automated because (I think?) it could be fully done remotely. But my guess is the most intelligent/good at abstract reasoning architects aren’t many multiples more productive than average architects.

  6. ^

     Although, compute to run experiments to work out which ideas are promising will be scarce, so this picture is a simplification, and having lots of costly-to-rule-out bad ideas is still problematic.

  7. ^

     For this factor to be important the government would likely need to play a key role in allocating AI to industries. This level of command and control may not happen

  8. Show all footnotes

39

0
0

Reactions

0
0

More posts like this

Comments3


Sorted by Click to highlight new comments since:

To be honest, I wouldn't personally recommend this Epoch article to people.

It has some strong analysis at points, but unfortunately, it's undermined by some poor choices of framing/focus that mean most readers will probably leave more confused than when they came.

• For a start, this article focuses almost purely on the economic impacts. However, if an AI cures cancer, the value to humanity will likely significantly exceed the economic value. Similarly, reducing emissions isn't going to directly lead to productivity growth in the short-term, but in the longer term, excessive emissions could damage the economy. Their criticism that Dario, Demis and Sam haven't engaged with their (purely economic) reframing is absurd.
• Similarly, the CEO's mostly aren't even attempting to address the question of how to apportion impacts between R&D or automation. They were trying to convince folks that the impact of AI could be massive (wherever it is coming from). So whilst there is a significant difference in worldviews with the CEO's being more bullish on automated R&D than the authors, the CEO's have been awkwardly shoehorned into the role of opposition in a subtly different debate.
• Their criticism of R&D focuses quite strong on a binary automated/non-automated frame. The potential of human and AI to work together on tasks is mostly neglected. They observe that R&D coding assistants haven't led to explosive R&D. However, automated coding only became really good recently. Many folks still aren't aware of how good these tools are or haven't adopted them because their organisation hasn't prioritised spending money on AI coding agents. The training pipeline or hiring practices haven't really adapted to the new reality yet. Organisations that better adopt these practices haven't had sufficient time to outcompete those that are slower. Unfortunately, their argument requires them to defend this point in order to go through. In other words, their entire argument is hanging by a thread.

So it was an interesting article. I learned things by reading it and reflecting on it, but I can't honestly recommend it to others as I had to waste a bit too much time trying to disentangle some of the poorly chosen frames.

I feel like the counterpoint here is that R&D is incredibly hard. In regular development, you have established methods of how to do things, established benchmarks of when things are going well, and a long period of testing to discover errors, flaws, and mistakes through trial and error. 

In R&D, you're trying to do things that nobody has ever done before, and simultaneously establish methods, benchmarks, and errors for that new method, which carries a ton of potential pitfalls. Also, nobody has ever done it before, so the AI is always inherently out-of-training to a much greater degree than in regular work. 

Yes, this seems right, hard to know which effect will dominate. I'm guessing you could assemble pretty useful training data of past R&D breakthroughs which might help, but that will only get you so far.

Curated and popular this week
 ·  · 5m read
 · 
Today, Forethought and I are releasing an essay series called Better Futures, here.[1] It’s been something like eight years in the making, so I’m pretty happy it’s finally out! It asks: when looking to the future, should we focus on surviving, or on flourishing? In practice at least, future-oriented altruists tend to focus on ensuring we survive (or are not permanently disempowered by some valueless AIs). But maybe we should focus on future flourishing, instead.  Why?  Well, even if we survive, we probably just get a future that’s a small fraction as good as it could have been. We could, instead, try to help guide society to be on track to a truly wonderful future.    That is, I think there’s more at stake when it comes to flourishing than when it comes to survival. So maybe that should be our main focus. The whole essay series is out today. But I’ll post summaries of each essay over the course of the next couple of weeks. And the first episode of Forethought’s video podcast is on the topic, and out now, too. The first essay is Introducing Better Futures: along with the supplement, it gives the basic case for focusing on trying to make the future wonderful, rather than just ensuring we get any ok future at all. It’s based on a simple two-factor model: that the value of the future is the product of our chance of “Surviving” and of the value of the future, if we do Survive, i.e. our “Flourishing”.  (“not-Surviving”, here, means anything that locks us into a near-0 value future in the near-term: extinction from a bio-catastrophe counts but if valueless superintelligence disempowers us without causing human extinction, that counts, too. I think this is how “existential catastrophe” is often used in practice.) The key thought is: maybe we’re closer to the “ceiling” on Survival than we are to the “ceiling” of Flourishing.  Most people (though not everyone) thinks we’re much more likely than not to Survive this century.  Metaculus puts *extinction* risk at about 4
 ·  · 6m read
 · 
This is a crosspost from my new Substack Power and Priorities where I’ll be posting about power grabs, AI governance strategy, and prioritization, as well as some more general thoughts on doing useful things.  Tl;dr I argue that maintaining nonpartisan norms on the EA Forum, in public communications by influential community members, and in funding decisions may be more costly than people realize. Lack of discussion in public means that people don’t take political issues as seriously as they should, research which depends on understanding the political situation doesn’t get done, and the community moves forward with a poor model of probably the most consequential actor in the world for any given cause area - the US government. Importantly, I don’t mean to say most community members shouldn’t maintain studious nonpartisanship! I merely want to argue that we should be aware of the downsides and do what we can to mitigate them.    Why nonpartisan norms in EA are a big deal Individual politicians (not naming names) are likely the most important single actors affecting the governance of AI. The same goes for most of the cause areas EAs care about. While many prominent EAs think political issues may be a top priority, and politics is discussed somewhat behind closed doors, there is almost no public discussion of politics. I argue the community’s lack of a public conversation about the likely impacts of these political actors and what to do in response to them creates large costs for how the community thinks about and addresses important issues (i.e. self-censorship matters actually). Some of these costs include:  * Perceived unimportance: I suspect a common, often subconscious, thought is, 'no prominent EAs are talking about politics publicly so it's probably not as big of a deal as it seems'. Lack of public conversation means social permission is never granted to discuss the issue as a top priority, it means the topic comes up less & so is thought about less, and i
 ·  · 4m read
 · 
Context: I’m a senior fellow at Conservation X Labs (CXL), and I’m seeking support as I attempt to establish a program on humane rodent fertility control in partnership with the Wild Animal Initiative (WAI) and the Botstiber Institute for Wildlife Fertility Control (BIWFC). CXL is a biodiversity conservation organization working in sustainable technologies, not an animal welfare organization. However, CXL leadership is interested in simultaneously promoting biodiversity conservation and animal welfare, and they are excited about the possibility of advancing applied research that make it possible to ethically limit rodent populations to protect biodiversity.  I think this represents the wild animal welfare community’s first realistic opportunity to bring conservation organizations into wild animal welfare work while securing substantial non-EA funding for welfare-improving interventions.  Background Rodenticides cause immense suffering to (likely) hundreds of millions of rats and mice annually through anticoagulation-induced death over several days, while causing significant non-target harm to other animals. In the conservation context, rodenticides are currently used in large-scale island rat and mouse eradications as a way of protecting endemic species. But these rodenticides kill lots of native species in addition to the mice and rats. So advancements in fertility control would be a benefit to both conservation- and welfare-focused stakeholders. CXL is a respected conservation organization with a track record of securing follow-on investments for technologies we support (see some numbers below). We are interested in co-organizing a "Big Think" workshop with WAI and BIWFC. The event will launch an open innovation program (e.g., a prize or a challenge process) to accelerate fertility control development. The program would specifically target island conservation applications where conservation groups are already motivated to replace rodenticides, but would likely