Self-studying to develop an inside-view model of AI alignment; co-studiers welcome! 2021-11-30T09:25:05.146Z


Comment by Vael Gates on Discussion with Eliezer Yudkowsky on AGI interventions · 2021-11-14T08:47:56.919Z · LW · GW

"Alpha Zero scales with more computing power, I think AlphaFold 2 scales with more computing power, Mu Zero scales with more computing power. Precisely because GPT-3 doesn't scale, I'd expect an AGI to look more like Mu Zero and particularly with respect to the fact that it has some way of scaling."

I thought GPT-3 was the canonical example of a model type that people are worried about will scale? (i.e. it's discussed in

Comment by Vael Gates on MichaelA's Shortform · 2021-09-23T01:00:33.598Z · LW · GW

Recently I was also trying to figure out what resources to send to an economist, and couldn't find a list that existed either! The list I came up with is subsumed by yours, except:
- Questions within Some AI Governance Research Ideas
- "Further Research" section within an OpenPhil 2021 report:
- The AI Objectives Institute just launched, and they may have questions in the future