The two missing core reasons why aligning at-least-partially superhuman AGI is hard

post by Joel Burget (joel-burget) · 2022-04-19T17:15:23.965Z · LW · GW · 2 comments

This is a question post.

Contents

2 comments

From Arbital's Mild Optimization page:

Mild optimization relates directly to one of the three core reasons why aligning at-least-partially superhuman AGI is hard - making very powerful optimization pressures flow through the system puts a lot of stress on its potential weaknesses and flaws.

I'm interested in this taxonomy of core reasons. Unfortunately this page doesn't specify the other two. What are they?

Also, this page is part of the AI alignment domain -- was it written by Eliezer? (surprisingly, "10 changes by 3 authors" is a link to edit and does not show author information or edit history)

Answers

2 comments

Comments sorted by top scores.

comment by Pattern · 2022-04-20T02:52:03.754Z · LW(p) · GW(p)
Three core reasons why

I'd have called this question, The Two missing core reasons why..., or Arbital's two missing core, etc.

Replies from: joel-burget
comment by Joel Burget (joel-burget) · 2022-04-20T14:14:32.647Z · LW(p) · GW(p)

Thanks Pattern -- I've taken your advice and updated the title.