Take the 2025 EA Forum Survey to help inform our strategy and prioritiesTake the survey
Hide table of contents

Transformative AI and Compute - A holistic approach - Part 3 out of 4

This is part three of the series Transformative AI and Compute - A holistic approach. You can find the sequence here and the summary here.

This work was conducted as part of Stanford’s Existential Risks Initiative (SERI) at the Center for International Security and Cooperation, Stanford University. Mentored by Ashwin Acharya (Center for Security and Emerging Technology (CSET)) and Michael Andregg (Fathom Radiant).

This post attempts to:

  1. Briefly outline the relevance of compute for AI Governance (Section 6).
  2. Conclude this report and discuss next steps (Section 7).

Epistemic Status

This article is Exploratory to My Best Guess. I've spent roughly 300 hours researching this piece and writing it up. I am not claiming completeness for any enumerations. Most lists are the result of things I learned on the way and then tried to categorize.

I have a background in Electrical Engineering with an emphasis on Computer Engineering and have done research in the field of ML optimizations for resource-constrained devices — working on the intersection of ML deployments and hardware optimization. I am more confident in my view on hardware engineering than in the macro interpretation of those trends for AI progress and timelines.

This piece was a research trial to test my prioritization, interest and fit for this topic. Instead of focusing on a single narrow question, this paper and research trial turned out to be more broad — therefore a holistic approach. In the future, I’m planning to work more focused on a narrow relevant research questions within this domain. Please reach out.

Views and mistakes are solely my own.

Previous Post: Forecasting Compute

You can find the previous post "Forecasting Compute [2/4]" here.

6. Compute Governance

Highlights

  • Compute is a unique AI governance node due to the required physical space, energy demand, and the concentrated supply chain. Those features make it a governable candidate.
  • Controlling and governing access to compute can be harnessed to achieve better AI safety outcomes, for instance restricting compute access to non-safety-aligned actors.
  • As compute becomes a dominant factor of costs at the frontier of AI research, it may start to resemble high-energy physics research, where a significant amount of the budget is spent on infrastructure (unlike previous trends of CS research where the equipment costs have been fairly low).

Lastly, I want to motivate the topic of compute governance as a subfield of AI governance and briefly highlight the unique aspect of compute governance.

Compute has three unique features which might make it more governable than other domains of AI governance (such as talent, ideas, and data) (Anderljung and Carlier 2021):

  1. Compute requires physical space for the computing hardware — football-field-sized supercomputer centers are the norm (Los Alamos National Laboratory 2013). Compared to software, this makes compute easier to track.
    • Additionally, compute is often highly centralized due to the dominance of cloud providers, such as Amazon Web Services (AWS), Google Cloud, and others. Moreover, current leading hardware, such as Google TPUs, is only available as a service. Consequently, this feature makes it more governable.
  2. The energy (and water demands). For running those supercomputers, massive amounts of energy and water for cooling are required (Los Alamos National Laboratory 2013).
  3. The supply chain of the semiconductor is highly concentrated, which could enable monitoring and governance (Khan 2021) — see “The Semiconductor Supply Chain” by CSET for more.

Second, according to my initial research and talking to people in the field of AI governance, there seems to be more of a consensus on what to do with compute regarding governance: restricting and regulating access to compute resources for less cautious actors.[1] This does not include a consensus on the concrete policies but at least in regards to the goal. Whereas for other aspects in the field of AI governance, there seems to be no clear consensus on which intermediate goals to pursue (see a discussion in this post).

6.1 Funding Allocation

Within this decade, we will and should see a switch in funding distribution at publicly funded AI research groups. Whereas AI and computer science (CS) research groups usually had relatively low overhead costs for equipment, this will change in the future to the increased need for spending more funding on compute to maintain state-of-the-art research. Those groups will become more like high-energy physics or biology research groups where considerable funding is being spent on infrastructure (e.g., equipment and hardware). If this does not happen, publicly funded groups will not be able to compete. We can already observe this compute divide (Ahmed and Wahed 2020).

6.2 Research Questions

For a list of research questions see some “Some AI Governance Research Ideas” (Anderljung and Carlier 2021). My research questions are listed in Appendix A, including some notes on compute governance-related points.

7. Conclusions

Highlights

  • In terms of published papers, the research on compute trends, compute spending, and algorithmic efficiency (the field of macro ML research) is minor and more work on this intersection could quickly improve our understanding.
  • The field is currently bottlenecked by available data on macro ML trends: total compute used to train a model is rarely published, nor is spending. With these it would be easier to estimate algorithmic efficiency and build better forecasting models.
  • The importance of compute also highlights the need for ML engineers working on AI safety to be able to deploy gigantic models.
    • Therefore, more people should consider becoming an AI hardware expert or working as an ML engineer at safety-aligned organizations and enabling their deployment success.
  • But also working on the intersection of technology and economics is relevant to inform spending and understanding of macro trends.
  • Research results in all of the mentioned fields could then be used to inform compute governance.

Compute is a substantial component of AI systems and has been a driver of their capabilities. Compared to data and algorithmic innovation, it provides a unique quantifiability that enables more efficient analysis and governance.

The effective available compute is mainly informed by the compute prices, the spending, and algorithmic improvements. Nonetheless, we should also explore the downsides of purely focusing on computational power and consider using metrics based on our understanding of the interconnect and memory capacity.

We have discussed components of hardware progress and discussed the recent trends such as Moore’s law, chip architectures, and hardware paradigms. Focusing on only one trend comes with significant shortcomings; instead, I suggest we inform our forecasts by combining such models. I would be especially excited to break down existing compute trends into hardware improvements and increased spending.

Limited research in the field of macro AI

My research is based on a small set of papers, whereas most focus on certain sub aspects. Overall, the research field of macro ML trends in used compute is, to my understanding, fairly small. Seeing more research efforts on compute trends and algorithmic innovation could be highly beneficial. This could lead to a better understanding of past trends, and forecasting future trends — for example, breaking down the trend into increased spending and hardware progress can give us some insights into potential upper limits.

Limited data for analyzing AI trends

Another limitation, and perhaps the cause of limited research, is that , there is also limited data available. Consequently, researchers first need to build the required dataset. I would be excited to see bigger datasets of compute requirements or experiments to measure algorithmic efficiency.

We share in this work our public ML progress dataset and a dataset using MLCommons training benchmarks (MLCommons 2021) for measuring the performance progress of modern AI hardware and ask others to share their insights and data.

ML deployment engineers

As the role of compute is significant for AI progress, there is a strong need for ML engineers who can efficiently deploy AI systems. This was also discussed by Olah in an 80’000 hours episode #107. Consequently, ML engineers should consider working at safety-aligned organizations and enable the deployment of gigantic models which are —ideally— reliable, interpretable and steerable.

Interdisciplinary research

An essential component for compute prices and spending are economic models — either based on spending, or the computing industry, such as the semiconductor industry. Interdisciplinary research on those questions could be of great benefit. Examples of such work are (Thompson et al. 2020; Thompson and Spanuth 2021).

I plan to work on aspects of this research in the future and would be especially interested in exploring collaboration or other synergies. Please reach out. The exact research questions are still to be determined.

Appendix A lists various research questions that I would be interested in exploring and also want others to explore.

Next Post: Compute Research Questions and Metrics

The appendix "Compute Research Questions and Metrics [4/4]" will attempt to:

  1. Provide a list of connected research questions (Appendix A).
  2. Present common compute metrics and discusses their caveats (Appendix B).
  3. Provide a list of Startups in the AI Hardware domain (Appendix C).

Acknowledgments

You can find the acknowledgments in the summary.

References

The references are listed in the summary.


  1. It seems reasonable and somewhat likely to me that we will be regulating and restricting the export of AI hardware even harsher and might classify it legally as weapons within the next decades. ↩︎

Comments3


Sorted by Click to highlight new comments since:

Really nice work, just got to reading it.

Those groups will become more like high-energy physics or biology research groups where considerable funding is being spent on infrastructure (e.g., equipment and hardware). If this does not happen, publicly funded groups will not be able to compete.

How certain are you about this? Your analogies for extremely costly research are both publicly funded groups, so it wouldn't seem too surprising to me if governments will start opening their pockets for research into what seems to have similar or greater scientific and public "excitement levels" than physics and biology.

I'm still holding the same view that (a) we will probably see a switch in funding distribution and (b) if this does not happen those groups won't be able to compete with SOTA models.

we will and should see a switch in funding distribution at publicly funded AI research groups

I would change my mind if we find more evidence towards algorithmic innovation being a stronger or the significant driver.

Some recent updates in regards to providing more funding for infrastructure include The National AI Research Cloud which is currently being investigated by the US government or Compute Canada.

Just realized that I misunderstood the original quote, yes, thanks, this makes total sense. 

Curated and popular this week
 ·  · 1m read
 · 
This morning I was looking into Switzerland's new animal welfare labelling law. I was going through the list of abuses that are now required to be documented on labels, and one of them made me do a double-take: "Frogs: Leg removal without anaesthesia."  This confused me. Why are we talking about anaesthesia? Shouldn't the frogs be dead before having their legs removed? It turns out the answer is no; standard industry practice is to cut their legs off while they are fully conscious. They remain alive and responsive for up to 15 minutes afterward. As far as I can tell, there are zero welfare regulations in any major producing country. The scientific evidence for frog sentience is robust - they have nociceptors, opioid receptors, demonstrate pain avoidance learning, and show cognitive abilities including spatial mapping and rule-based learning.  It's hard to find data on the scale of this issue, but estimates put the order of magnitude at billions of frogs annually. I could not find any organisations working directly on frog welfare interventions.  Here are the organizations I found that come closest: * Animal Welfare Institute has documented the issue and published reports, but their focus appears more on the ecological impact and population decline rather than welfare reforms * PETA has conducted investigations and released footage, but their approach is typically to advocate for complete elimination of the practice rather than welfare improvements * Pro Wildlife, Defenders of Wildlife focus on conservation and sustainability rather than welfare standards This issue seems tractable. There is scientific research on humane euthanasia methods for amphibians, but this research is primarily for laboratory settings rather than commercial operations. The EU imports the majority of traded frog legs through just a few countries such as Indonesia and Vietnam, creating clear policy leverage points. A major retailer (Carrefour) just stopped selling frog legs after welfar
 ·  · 10m read
 · 
This is a cross post written by Andy Masley, not me. I found it really interesting and wanted to see what EAs/rationalists thought of his arguments.  This post was inspired by similar posts by Tyler Cowen and Fergus McCullough. My argument is that while most drinkers are unlikely to be harmed by alcohol, alcohol is drastically harming so many people that we should denormalize alcohol and avoid funding the alcohol industry, and the best way to do that is to stop drinking. This post is not meant to be an objective cost-benefit analysis of alcohol. I may be missing hard-to-measure benefits of alcohol for individuals and societies. My goal here is to highlight specific blindspots a lot of people have to the negative impacts of alcohol, which personally convinced me to stop drinking, but I do not want to imply that this is a fully objective analysis. It seems very hard to create a true cost-benefit analysis, so we each have to make decisions about alcohol given limited information. I’ve never had problems with alcohol. It’s been a fun part of my life and my friends’ lives. I never expected to stop drinking or to write this post. Before I read more about it, I thought of alcohol like junk food: something fun that does not harm most people, but that a few people are moderately harmed by. I thought of alcoholism, like overeating junk food, as a problem of personal responsibility: it’s the addict’s job (along with their friends, family, and doctors) to fix it, rather than the job of everyday consumers. Now I think of alcohol more like tobacco: many people use it without harming themselves, but so many people are being drastically harmed by it (especially and disproportionately the most vulnerable people in society) that everyone has a responsibility to denormalize it. You are not likely to be harmed by alcohol. The average drinker probably suffers few if any negative effects. My argument is about how our collective decision to drink affects other people. This post is not
 ·  · 5m read
 · 
Today, Forethought and I are releasing an essay series called Better Futures, here.[1] It’s been something like eight years in the making, so I’m pretty happy it’s finally out! It asks: when looking to the future, should we focus on surviving, or on flourishing? In practice at least, future-oriented altruists tend to focus on ensuring we survive (or are not permanently disempowered by some valueless AIs). But maybe we should focus on future flourishing, instead.  Why?  Well, even if we survive, we probably just get a future that’s a small fraction as good as it could have been. We could, instead, try to help guide society to be on track to a truly wonderful future.    That is, I think there’s more at stake when it comes to flourishing than when it comes to survival. So maybe that should be our main focus. The whole essay series is out today. But I’ll post summaries of each essay over the course of the next couple of weeks. And the first episode of Forethought’s video podcast is on the topic, and out now, too. The first essay is Introducing Better Futures: along with the supplement, it gives the basic case for focusing on trying to make the future wonderful, rather than just ensuring we get any ok future at all. It’s based on a simple two-factor model: that the value of the future is the product of our chance of “Surviving” and of the value of the future, if we do Survive, i.e. our “Flourishing”.  (“not-Surviving”, here, means anything that locks us into a near-0 value future in the near-term: extinction from a bio-catastrophe counts but if valueless superintelligence disempowers us without causing human extinction, that counts, too. I think this is how “existential catastrophe” is often used in practice.) The key thought is: maybe we’re closer to the “ceiling” on Survival than we are to the “ceiling” of Flourishing.  Most people (though not everyone) thinks we’re much more likely than not to Survive this century.  Metaculus puts *extinction* risk at about 4