HLAI 2018 Field Reportpost by G Gordon Worley III (gworley) · 2018-08-29T00:11:26.106Z · score: 51 (21 votes) · LW · GW · 12 comments
I spent the past week in Prague (natively called Praha, but known widely as Prague via the usual mechanisms others decide your own names) at the Human Level AI multi-conference held between the AGI, BICA, the NeSY conferences. It also featured the Future of AI track, which was my proximate reason for attending to speak on the Solving the AI Race panel thanks to the selection of my submission to GoodAI's Solving the AI Race challenge as one of the finalists. I enjoyed the conference tremendously and the people I met there even more, and throughout the four days I found myself noticing things I felt were worth sharing more widely, hence this field report.
I arrived on Tuesday and stayed at the home of a fellow EA (they can out themselves in the comments if they like, but I won't do it here since I couldn't get in touch with them in time to get their permission to disclose their personage). If you didn't already know, Prague is rapidly becoming a hub for effective altruists in Europe, and having visited it's easy to see why: the city is beautiful, most people speak enough English that communication is easy for everyone, the food is delicious (there's lots of vegan food even though traditional Czech cuisine is literally meat and potatoes), and it's easily accessible to everyone on the continent. Some of the recent events held in Prague include the Human-aligned AI Summer School and the upcoming Prague AI safety camp.
Wednesday was the first day of the conference, and I honestly wasn't quite sure what to expect, but Ben Goertzel did a great job of setting the tone with his introductory keynote that made clear the focus was on exploring ideas related to building AGI. We then dove in to talk after talk for the next 4 days, each one considering an idea that might lie on the path to enabling general intelligence for machines, and we doubled-up on the third day with a Future of AI program that considered AI policy and ethics issues. I took a lot of notes, but being a safety researcher I'm hesitant to share them because that feels like a hazard that might accelerate AGI development, so instead I'll share a few high-level insights that I think summarize what I learned without saying anything with more than a very small (let's say < 5%) chance of giving someone a dangerous inspiration they wouldn't have had anyway or as soon.
- Progress towards AGI is not amorphous; it's being made through small starts by individuals with faces and names. I knew this intellectually, but it was different to live it—shaking people's hands, swapping stories, and hearing them express their hopes and doubts that don't appear in published work because it doesn't fit the academic form. That said, I still find their work threatening, but now I know the threat on a human level.
- Thankfully we are further away than it often feels like if you focus on the success of deep learning systems. Nearly everyone at the conference agreed that deep learning in its current form is missing some key ingredients to get to AGI. Again, I won't repeat here what those are, but the good news is that deep learning appears to not be on the path to AGI even if it can produce non-general superintelligence, and AGI requires new paradigms in machine intelligence that we either haven't discovered or don't yet know how to make work well in practice.
- From talking to people I have a stronger model of why they work on capabilities and not safety or, as is more often the case, think safety concerns do not require extreme caution, instead necessitating only normal levels of caution as is the case in most engineering disciplines. My read from dozens of similar interactions with researchers is that this attitude reflects both a lack of deep reflection on safety and a difficulty in updating to an idea that would break the momentum of their life's work.
- But not everyone was like this. For example, once people figured out I was a "safety guy", one person came up to me and asked what he should do because he has an idea he thinks would move AI in a dangerous direction and didn't know if he should publish anyway, keep it a secret, or do something else. I recommended that he reach out to MIRI, and I hope he will, but it exposed to me that even when a capabilities researcher notices their work is going in a dangerous direction they don't have a standard script for what to do about it.
- On a different note, EA continues to have perception problems. I talked to two people who said things that indicated they lean EA, asked them about if they identified that way, and then they told me they didn't because they associate EA with Singer-style act utilitarianism and self-imposed poverty through maximizing donated income. I let them know the EA tent is much bigger than that, but clearly EA evangelists have more work to do in this direction.
To summarize, I think the main takeaway is that we, being the sorts of persons who read LessWrong, live in a bubble where we know AGI is dangerous, and outside that bubble people still don't know or have confused ideas about how it's dangerous, even among the group of people weird enough to work on AGI instead of more academically respectable, narrow AI. That's really scary, because the people outside the bubble also include those affecting public policy and making business decisions, and they lack the perspective we share about the dangers of both narrow and general AI. Luckily, this points towards two opportunities we can work on now to mitigate the risks of AGI:
- Normalize thinking about AI safety. Right now it would be a major improvement if we could move the field of AI research to be on par with the way biomedical researchers think about risks for near-term, narrow applications of their research, let alone getting everyone thinking about the existential risks of AGI. I think most of this work needs to happen on a 1-to-1, human level right now, pushing individual researchers towards more safety-focus and reshaping the culture of the AI research community. For this reason I think it's extremely important that I and others make an effort to
- attend AI capabilities conferences,
- form personal relationships with AI researchers,
- and encourage them to take safety seriously.
- Establish and publicize a "sink" for dangerous AI research. When people have an idea they think is dangerous (which assumes to some extent we succeed at the previous objective, but as I mentioned it already comes up now), they need a default script for what to do. Cybersecurity and biomedical research have standard approaches, and although I don't think their approaches will work the same for AGI, we can use them as models for designing a standard. The sink source should then be owned by a team seen as extremely responsible, reliable, and committed to safety above all else. I recommend FHI or MIRI (or both!) take on that role. The sub-actions of this work are to
- design and establish a process,
- find it a home,
- publicize its use,
- and continually demonstrate its effectiveness—especially to capabilities researchers who might have dangerous ideas—so it remains salient.
These interventions are more than I can take on myself, and I don't believe I have a comparative advantage to execute on them, so I need your help if you're interested, i.e. if you've been thinking about doing more with AI safety, know that there is concrete work you can do now other than technical work on alignment. For myself I've set an intention that when I attend HLAI 2020 we'll have moved at least half-way towards achieving these goals, so I'll be working on them as best I can, but we're not going to get there if I have to do it alone. If you'd like to coordinate on these objectives feel free to start in the comments below or reach out to me personally and we can talk more.
I feel like I've only just scratched the surface of my time at HLAI 2018 in this report, and I think it will take a while to process everything I learned and follow-up with everyone I talked to there. But if I had to give my impression of the conference in a tweet it would be this: we've come a long way since 2014, and I'm very pleased with the progress (Superintelligence ⇒ Puerto Rico ⇒ Asilomar), but we have even further to go, so let's get to work!
Comments sorted by top scores.