Compute Governance and Conclusions - Transformative AI and Compute [3/4]
post by lennart · 2021-10-14T08:23:12.156Z · LW · GW · 0 commentsContents
Epistemic Status Previous Post: Forecasting Compute 6. Compute Governance 6.1 Funding Allocation 6.2 Research Questions 7. Conclusions Limited research in the field of macro AI Limited data for analyzing AI trends ML deployment engineers Interdisciplinary research Next Post: Compute Research Questions and Metrics Acknowledgments References None No comments
Cross-posted here on the EA Forum [EA · GW].
Transformative AI and Compute - A holistic approach - Part 3 out of 4
This is part two of the series Transformative AI and Compute - A holistic approach. You can find the sequence here [? · GW] and the summary here [LW · GW].
This work was conducted as part of Stanford’s Existential Risks Initiative (SERI) at the Center for International Security and Cooperation, Stanford University. Mentored by Ashwin Acharya (Center for Security and Emerging Technology (CSET)) and Michael Andregg (Fathom Radiant).
This post attempts to:
- Briefly outline the relevance of compute for AI Governance (Section 6 [LW · GW]).
- Conclude this report and discuss next steps (Section 7 [LW · GW]).
Epistemic Status
This article is Exploratory to My Best Guess. I've spent roughly 300 hours researching this piece and writing it up. I am not claiming completeness for any enumerations. Most lists are the result of things I learned on the way and then tried to categorize.
I have a background in Electrical Engineering with an emphasis on Computer Engineering and have done research in the field of ML optimizations for resource-constrained devices — working on the intersection of ML deployments and hardware optimization. I am more confident in my view on hardware engineering than in the macro interpretation of those trends for AI progress and timelines.
This piece was a research trial to test my prioritization, interest and fit for this topic. Instead of focusing on a single narrow question, this paper and research trial turned out to be more broad — therefore a holistic approach. In the future, I’m planning to work more focused on a narrow relevant research questions within this domain. Please reach out.
Views and mistakes are solely my own.
Previous Post: Forecasting Compute
You can find the previous post "Forecasting Compute [2/4]" here [LW · GW].
6. Compute Governance
Highlights
- Compute is a unique AI governance node due to the required physical space, energy demand, and the concentrated supply chain. Those features make it a governable candidate.
- Controlling and governing access to compute can be harnessed to achieve better AI safety outcomes, for instance restricting compute access to non-safety-aligned actors.
- As compute becomes a dominant factor of costs at the frontier of AI research, it may start to resemble high-energy physics research, where a significant amount of the budget is spent on infrastructure (unlike previous trends of CS research where the equipment costs have been fairly low).
Lastly, I want to motivate the topic of compute governance as a subfield of AI governance and briefly highlight the unique aspect of compute governance.
Compute has three unique features which might make it more governable than other domains of AI governance (such as talent, ideas, and data) (Anderljung and Carlier 2021):
- Compute requires physical space for the computing hardware — football-field-sized supercomputer centers are the norm (Los Alamos National Laboratory 2013). Compared to software, this makes compute easier to track.
- Additionally, compute is often highly centralized due to the dominance of cloud providers, such as Amazon Web Services (AWS), Google Cloud, and others. Moreover, current leading hardware, such as Google TPUs, is only available as a service. Consequently, this feature makes it more governable.
- The energy (and water demands). For running those supercomputers, massive amounts of energy and water for cooling are required (Los Alamos National Laboratory 2013).
- The supply chain of the semiconductor is highly concentrated, which could enable monitoring and governance (Khan 2021) — see “The Semiconductor Supply Chain” by CSET for more.
Second, according to my initial research and talking to people in the field of AI governance, there seems to be more of a consensus on what to do with compute regarding governance: restricting and regulating access to compute resources for less cautious actors.[1] This does not include a consensus on the concrete policies but at least in regards to the goal. Whereas for other aspects in the field of AI governance, there seems to be no clear consensus on which intermediate goals to pursue (see a discussion in this post [EA · GW]).
6.1 Funding Allocation
Within this decade, we will and should see a switch in funding distribution at publicly funded AI research groups. Whereas AI and computer science (CS) research groups usually had relatively low overhead costs for equipment, this will change in the future to the increased need for spending more funding on compute to maintain state-of-the-art research. Those groups will become more like high-energy physics or biology research groups where considerable funding is being spent on infrastructure (e.g., equipment and hardware). If this does not happen, publicly funded groups will not be able to compete. We can already observe this compute divide (Ahmed and Wahed 2020).
6.2 Research Questions
For a list of research questions see some “Some AI Governance Research Ideas” (Anderljung and Carlier 2021). My research questions are listed in Appendix A [LW · GW], including some notes on compute governance-related points.
7. Conclusions
Highlights
- In terms of published papers, the research on compute trends, compute spending, and algorithmic efficiency (the field of macro ML research) is minor and more work on this intersection could quickly improve our understanding.
- The field is currently bottlenecked by available data on macro ML trends: total compute used to train a model is rarely published, nor is spending. With these it would be easier to estimate algorithmic efficiency and build better forecasting models.
- The importance of compute also highlights the need for ML engineers working on AI safety to be able to deploy gigantic models.
- Therefore, more people should consider becoming an AI hardware expert [EA · GW] or working as an ML engineer at safety-aligned organizations and enabling their deployment success.
- But also working on the intersection of technology and economics is relevant to inform spending and understanding of macro trends.
- Research results in all of the mentioned fields could then be used to inform compute governance.
Compute is a substantial component of AI systems and has been a driver of their capabilities. Compared to data and algorithmic innovation, it provides a unique quantifiability that enables more efficient analysis and governance.
The effective available compute is mainly informed by the compute prices, the spending, and algorithmic improvements. Nonetheless, we should also explore the downsides of purely focusing on computational power and consider using metrics based on our understanding of the interconnect and memory capacity.
We have discussed components of hardware progress and discussed the recent trends such as Moore’s law, chip architectures, and hardware paradigms. Focusing on only one trend comes with significant shortcomings; instead, I suggest we inform our forecasts by combining such models. I would be especially excited to break down existing compute trends into hardware improvements and increased spending.
Limited research in the field of macro AI
My research is based on a small set of papers, whereas most focus on certain sub aspects. Overall, the research field of macro ML trends in used compute is, to my understanding, fairly small. Seeing more research efforts on compute trends and algorithmic innovation could be highly beneficial. This could lead to a better understanding of past trends, and forecasting future trends — for example, breaking down the trend into increased spending and hardware progress can give us some insights into potential upper limits.
Limited data for analyzing AI trends
Another limitation, and perhaps the cause of limited research, is that , there is also limited data available. Consequently, researchers first need to build the required dataset. I would be excited to see bigger datasets of compute requirements or experiments to measure algorithmic efficiency.
We share in this work our public ML progress dataset and a dataset using MLCommons training benchmarks (MLCommons 2021) for measuring the performance progress of modern AI hardware and ask others to share their insights and data.
ML deployment engineers
As the role of compute is significant for AI progress, there is a strong need for ML engineers who can efficiently deploy AI systems. This was also discussed by Olah in an 80’000 hours episode #107. Consequently, ML engineers should consider working at safety-aligned organizations and enable the deployment of gigantic models which are —ideally— reliable, interpretable and steerable.
Interdisciplinary research
An essential component for compute prices and spending are economic models — either based on spending, or the computing industry, such as the semiconductor industry. Interdisciplinary research on those questions could be of great benefit. Examples of such work are (Thompson et al. 2020; Thompson and Spanuth 2021).
I plan to work on aspects of this research in the future and would be especially interested in exploring collaboration or other synergies. Please reach out. The exact research questions are still to be determined.
Appendix A [LW · GW] lists various research questions that I would be interested in exploring and also want others to explore.
Next Post: Compute Research Questions and Metrics
The appendix "Compute Research Questions and Metrics [4/4] [LW · GW]" will attempt to:
- Provide a list of connected research questions (Appendix A [LW · GW]).
- Present common compute metrics and discusses their caveats (Appendix B [LW · GW]).
- Provide a list of Startups in the AI Hardware domain (Appendix C [LW · GW]).
Acknowledgments
You can find the acknowledgments in the summary [LW · GW].
References
The references are listed in the summary [LW · GW].
It seems reasonable and somewhat likely to me that we will be regulating and restricting the export of AI hardware even harsher and might classify it legally as weapons within the next decades. ↩︎
0 comments
Comments sorted by top scores.