How to get into AI safety research
post by Stuart_Armstrong · 2022-05-18T18:05:06.526Z · LW · GW · 7 commentsContents
7 comments
Recently, I had a conversation with someone from a math background, asking how they could get into AI safety research. Based on my own path from mathematics to AI alignment, I recommended the following sources. It may prove useful to others contemplating a similar change in career:
- Superintelligence by Nick Bostrom. It condenses all the main arguments for the power and the risk of AI, and gives a framework in which to think of the challenges and possibilities.
- Sutton and Barto's Book: Reinforcement Learning: An Introduction. This gives the very basics of what ML researchers actually do all day, and is important for understanding more advanced concepts. It gives (most of) the vocabulary to understand what ML and AI papers are talking about.
- Gödel without too many tears. This is how I managed to really grok logic and the completeness/incompleteness theorems. Important for understanding many of MIRI's and LessWrong's approaches to AI and decision theory.
- Safely Interruptible agents. It feels bad to recommend one of my own papers, but I think this is an excellent example of bouncing between ML concepts and alignment concepts to make some traditional systems interruptible (so that we can shut them down without them resisting the shutdown).
- Alignment for Advanced Machine Learning Systems. Helps give an overall perspective on different alignment methods, and some understanding of MIRI's view on the subject (for a deeper understanding, I recommend diving into MIRI's or Eliezer's publications/writings).
You mileage may vary, but these are the sources that I would recommend. And I encourage you to post any sources you'd recommend, in the comments.
7 comments
Comments sorted by top scores.
comment by JanB (JanBrauner) · 2022-05-19T07:33:42.415Z · LW(p) · GW(p)
I guess I'd recommend the AGI safety fundamentals course: https://www.eacambridge.org/technical-alignment-curriculum
On Stuart's list: I think this list might be suitable for some types of conceptual alignment research. But you'd certainly want to read more ML for other types of alignment research.
comment by Gunnar_Zarncke · 2022-05-19T21:39:55.162Z · LW(p) · GW(p)
This is nice from a "what do I need to study" perspective, but it does help less with the "how do I pay the bills" perspective. Do you have pointers there too?
Replies from: conor-sullivan↑ comment by Lone Pine (conor-sullivan) · 2022-05-19T22:21:55.427Z · LW(p) · GW(p)
AI Safety Support
https://www.aisafetysupport.org/resources/career-coaching
Replies from: Gunnar_Zarncke↑ comment by Gunnar_Zarncke · 2022-05-19T22:59:09.559Z · LW(p) · GW(p)
Thank you! I have scheduled a call.
comment by Joel Burget (joel-burget) · 2022-06-03T21:10:55.475Z · LW(p) · GW(p)
Thank you for mentioning Gödel Without Too Many Tears, which I bought it based on this recommendation. It's a lovely little book. I didn't expect to it to be nearly so engrossing.
Replies from: Stuart_Armstrong↑ comment by Stuart_Armstrong · 2022-06-05T03:09:35.959Z · LW(p) · GW(p)
Glad your liked it :-)
comment by Jsevillamol · 2022-05-19T22:10:59.522Z · LW(p) · GW(p)
I also found this thread of math topics on AI safety helpful.
https://forum.effectivealtruism.org/posts/d7fJLQz2QaDNbbWxJ/what-are-the-coolest-topics-in-ai-safety-to-a-hopelessly [EA · GW]