AI Safety and Neighboring Communities: A Quick-Start Guide, as of Summer 2022

post by Sam Bowman (sbowman) · 2022-09-01T19:15:40.713Z · LW · GW · 2 comments

Contents

  AI Safety/AI Alignment/AGI Safety/AI Existential Safety/AI X-Risk
  Effective Altruism/EA
  Longtermism
  The Rationalist Subculture/The LessWrong Crowd/Berkeley-Style Rationalism/The Rats
  AGI Optimism
  AI Ethics/Responsible AI/The FAccT Community
  (Long-Term) AI Governance
    Acknowledgments
None
2 comments

Getting into AI safety involves working with a mix of communities, subcultures, goals, and ideologies that you may not have encountered in the context of mainstream AI technical research. This document attempts to briefly map these out for newcomers.

This is inevitably going to be biased by what sides of these communities I (Sam) have encountered, and it will quickly become dated. I expect it will still be a useful resource for some people anyhow, at least in the short term.

AI Safety/AI Alignment/AGI Safety/AI Existential Safety/AI X-Risk

The research project of ensuring that future AI progress doesn’t yield civilization-endingly catastrophic results.

 

Effective Altruism/EA

The research project and social movement of doing as much good as possible with limited resources.

 

Longtermism

The ethical principle that the consequences of our actions on other people matter equally wherever and whenever those consequences are felt. Because there are a potentially huge number of future people we could influence by our choices, this says that considering our influence on the longer-term future should be a central part of ethical decision-making.

 

The Rationalist Subculture/The LessWrong Crowd/Berkeley-Style Rationalism/The Rats

A distinctive social group focused on using reason and science as thoroughly and deeply as possible in everyday life and important life decisions.

 

AGI Optimism

The view that building (aligned) AGI will lead to a post-scarcity, galaxy-spanning, pluralist utopia and would be humanity’s greatest achievement.

 

AI Ethics/Responsible AI/The FAccT Community

The research and political project of minimizing the harms of current and near-future AI/ML technology and of ensuring that any benefits from such technology are shared broadly.

 

(Long-Term) AI Governance

The project of developing institutions and policies within present-day governments to help increase the chances that AI progress goes well.


Acknowledgments

Thanks to Alex Tamkin, Jared Kaplan, Neel Nanda, Leo Gao, Fazl Barez, Owain Evans, Beth Barnes, and Rohin Shah for comments on a previous version of this.

2 comments

Comments sorted by top scores.

comment by Nathan Helm-Burger (nathan-helm-burger) · 2022-09-02T17:55:59.000Z · LW(p) · GW(p)

Super handy seeming intro for newcomers.

I recommend adding Jade Leung to your list of governance people.

As for the list of AI safety people, I'd like to add that there are some people who've written interesting and much discussed content that it would be worth having some familiarity with.

John Wentworth

Steven Byrnes

Vanessa Kosoy

And personally I'm quite excited about the school of thought developing under the 'Shard theory' banner.

For shard theory info:

https://www.lesswrong.com/posts/xqkGmfikqapbJ2YMj/shard-theory-an-overview [LW · GW

https://www.alignmentforum.org/posts/vJFdjigzmcXMhNTsx/simulators [AF · GW

Replies from: sbowman
comment by Sam Bowman (sbowman) · 2022-09-04T21:51:41.519Z · LW(p) · GW(p)

Thanks! I'll keep my opinionated/specific overview of the alignment community, but I know governance less well, so I'm happy to defer there.