Transformative AI and Compute [Summary]

post by lennart · 2021-09-26T11:41:44.731Z · LW · GW · 0 comments

Contents

  0. Executive Summary
    Abstract
    Epistemic Status
    Highlights per Section
      1. Compute
      2. Compute in AI Systems
      3. Compute and AI Alignment
      4. Forecasting Compute
      5. Better Compute Forecasts
      6. Compute Governance
      7. Conclusions
  Acknowledgments
  References
None
No comments

Cross-posted here on the EA Forum [EA · GW].

This is the summary of the series Transformative AI and Compute - A holistic approach. You can find the sequence here [? · GW] and the links to the posts below:

  1. Compute [1/4] [? · GW]
  2. Forecasting Compute [2/4] [LW · GW]
  3. Compute Governance and Conclusions [3/4] [LW · GW]
  4. Compute Research Questions and Metrics [4/4] [LW · GW]

0. Executive Summary

This series attempts to:

  1. Introduce a simplified model of computing which serves as a foundational concept (Part 1 - Section 1 [? · GW]).
  2. Discuss the role of compute for AI systems (Part 1 - Section 2 [? · GW]).
  3. Explore the connection of compute trends and more capable AI systems over time (Part 1 - Section 3 [? · GW]).
  4. Discuss the compute component in forecasting efforts on transformative AI timelines (Part 2 - Section 4 [LW · GW])
  5. Propose ideas for better compute forecasts (Part 2 - Section 5 [LW · GW]).
  6. Briefly outline the relevance of compute for AI Governance (Part 3 - Section 6 [LW · GW]).
  7. Conclude this report and discuss next steps (Section 7 [LW · GW]).
  8. Provide a list of connected research questions (Appendix A [LW · GW]).
  9. Present common compute metrics and discusses their caveats (Appendix B [LW · GW]).
  10. Provide a list of Startups in the AI Hardware domain (Appendix C [LW · GW]).

Abstract

Modern progress in AI systems has been driven and enabled mainly by acquiring more computational resources. AI systems rely on computation-intensive training runs — they require massive amounts of compute.

Learning about the compute requirements for training existing AI systems and their capabilities allows us to get a more nuanced understanding and take appropriate action within the technical and governance domain to enable a safe development of potential transformative AI systems.

To understand the role of compute, I decided to (a) do a literature review, (b) update existing work with new data, (c) investigate the role of compute for timelines, and lastly, (d) explore concepts to enhance our analysis and forecasting efforts.

In this piece, I present a brief analysis of AI systems’ compute requirements and capabilities, explore compute’s role for transformative AI timelines, and lastly, discuss the compute governance domain.

I find that compute, next to data and algorithmic innovation, is a crucial contributor to the recent performance of AI systems. We identify a doubling time of 6.2 months for the compute requirements of the final training run of state-of-the-art AI systems from 2012 to the present.
Next to more powerful hardware components, the spending on AI systems and the algorithmic innovation are other factors that inform the amount of effective compute available — which itself is a component for forecasting models on transformative AI.

Therefore, as compute is a significant component and driver of AI systems’ capabilities, understanding the developments of the past and forecasting future results is essential. Compared to the other components, the quantifiable nature of compute makes it an exciting aspect for forecasting efforts and the safe development of AI systems.

I consequently recommend additional investigations in highlighted components of compute, especially AI hardware. As compute forecasting and regulations require an in-depth understanding of hardware, hardware spending, the semiconductor industry, and much more, we recommend an interdisciplinary effort to inform compute trends interpretations and forecasts. Those insights can then be used to inform policymaking, and potentially regulate access to compute.

Epistemic Status

This article is Exploratory to My Best Guess. I've spent roughly 300 hours researching this piece and writing it up. I am not claiming completeness for any enumerations. Most lists are the result of things I learned on the way and then tried to categorize.

I have a background in Electrical Engineering with an emphasis on Computer Engineering and have done research in the field of ML optimizations for resource-constrained devices — working on the intersection of ML deployments and hardware optimization. I am more confident in my view on hardware engineering than in the macro interpretation of those trends for AI progress and timelines.

This piece was a research trial to test my prioritization, interest, and fit for this topic. Instead of focusing on a single narrow question, this paper and research trial turned out to be more broad — therefore a holistic approach. In the future, I’m planning to work more focused on a narrow relevant research question within this domain. Please reach out.

Views and mistakes are solely my own.

Highlights per Section

1. Compute [? · GW]

2. Compute in AI Systems [? · GW]

3. Compute and AI Alignment [? · GW]

4. Forecasting Compute [LW · GW]

5. Better Compute Forecasts [LW · GW]

6. Compute Governance [LW · GW]

7. Conclusions [LW · GW]

Acknowledgments

This work was supported and conducted as a summer fellowship at the Stanford Existential Risks Initiative (SERI). Their support is gratefully acknowledged. I am thankful for joining this program and would like to thank the organizers for enabling this, and the other fellows for the insightful discussions.

I am incredibly grateful to Ashwin Acharya and Michael Andregg for their mentoring throughout the project. Michaels thoughts on AI hardware nudged me to reconsider my current research interest and learn more about AI and compute. Ashwin for bouncing off ideas, the wealth of expertise in the domain, and helping me put things into the proper context. Thanks for the input! I was looking forward to every meeting and the thought-provoking discussions.

Thanks to the Swiss Existential Risk Initiative (CHERI) for providing the social infrastructure during my project. Having the opportunity to organize such an initiative in a fantastic team and being accompanied by motivated young researchers is astonishing.

I would like to express my thanks to Jaime Sevilla, Charlie Giattino, Will Hunt, Markus Anderljung, and Christopher Phenicie for your input and discussing ideas.

Thanks to Jaime Sevilla, Jeffrey Ohl, Christopher Phenicie, Aaron Gertler, and Kwan Yee Ng for providing feedback on this piece.

References


  1. Transformative AI, as defined by Open Philanthropy in this blogpost: “Roughly and conceptually, transformative AI is AI that precipitates a transition comparable to (or more significant than) the agricultural or industrial revolution.↩︎

0 comments

Comments sorted by top scores.