Stopping dangerous AI: Ideal lab behavior

post by Zach Stein-Perlman · 2023-05-09T21:00:19.505Z · LW · GW · 0 comments

Contents

No comments

Disclaimer: this post doesn't have the answers. Moreover, it's an unfinished draft. Hopefully a future version will be valuable, but that will only occur if I revise/rewrite it. For now you're better off reading sources linked from Ideas for AI labs: Reading list [LW · GW] and Slowing AI: Reading list [? · GW].

Why "stopping dangerous AI" rather than "slowing AI" here? It's more appropriate for actors' actions. Just "slowing" is a precise/continuous variable and what strategists think about while "stopping" is a natural/simple thought and what actors naturally aim for. Note that in the context of slowing, "AI" always meant "dangerous AI."

Blackbox AI safety (and how to do good with powerful AI) to focus on the speed of AI progress. What should leading labs do to facilitate a safe transition to a world with powerful AI? What would I do and plan if I controlled a leading lab?[1]

It would be better if progress toward dangerous AI capabilities was slower, all else equal. (This will become much more true near the end.) Or: it would be better if dangerous AI capabilities appear later, all else equal. (And slowing near the end is particularly important.)

 

Ways labs can slow AI:

(There are lots of other things labs should do for safety– this is just about slowing AI.)

(Note that what it would be best for an actor to do and what we should try to nudge that actor to do are different, because of not just tractability but also optimal change on the margin not necessarily being in the direction of optimal policy. For example, perhaps if you have lots of control over an unsafe lab you should focus on making it safer, while if you have little control you should focus on differentially slowing it.)

Leading labs are doing some important good things. DeepMind, Anthropic, and OpenAI seem to be avoiding publishing research-on-scary-paths to varying degrees. OpenAI's stop-and-assist clause is a first step toward coordination between labs to slow down near the end. OpenAI seems to be supporting evals. I don't know the details of labs' advocacy to the US government but I think it's mostly good.

Meta: figuring out what labs could do is pretty different from choosing between different well-specified things labs could do.

[2]

  1. ^

    Assuming I knew I would remain in control of the lab. Note that what we should nudge labs to do on the margin in the real world is not necessarily what they should do to act optimally.

  2. ^

    List considerations. Or, model: pretend there's a single critical deployment;* what determines whether the critical deployment goes well? Quality-adjusted alignment research done + alignment tax paid on that particular model?

    So:

    - [how much slowing occurs near the end] [combination of safety-conscious lead time and coordination among leading actors]

    - [doing alignment research and probably improving public/expert/policymaker opinion and facilitating good public policy is good– details mostly omitted here? And preparing to do good with powerful AI]

    - [note that some AI is on the dangerous-path and some isn't] 

    Desiderata/goals:

    - Minimize sharing information or tools that accelerate incautious actors

    - Prepare to slow down near the end?

    - [what public policy should they support? Export controls? Compute monitoring and evals regime?]

    - Coordinate / prepare to coordinate / tell others what you believe

    - Stop and assist?

    - Transparency?

     

    *On AI-agent-takeover threat models, this is a simplification but often a reasonable one; on WFLL threat models, it may be unreasonable.

0 comments

Comments sorted by top scores.