Reframing misaligned AGI's: well-intentioned non-neurotypical assistants

post by zhukeepa · 2018-04-01T01:22:36.993Z · LW · GW · 14 comments

Contents

14 comments

I think when people imagine misaligned AGI's, they tend to imagine a superintelligent agent optimizing for something other than human values (e.g. paperclips, or a generic reward signal), and mentally picture them as adversarial or malevolent. I think this visualization isn't as applicable for AGI's trained to optimize for human approval, like act-based agents, and I'd like to present one that is.

If you've ever employed someone or had a personal assistant, you might know that the following things are consistent:

Suppose you were considering hiring a personal assistant, and you knew a few things about it:

You might consider hiring this assistant and trying really, really hard to communicate to it exactly what you want. It seems like a way better idea to just not hire this assistant. Actually, you’d probably want to run for the hills if you were forced to. Some specific failure modes you might envision:

As an illustration of the above, imagine giving an eager, brilliant, extremely non-neurotypical friend free rein to help you find a romantic partner (e.g. helping you write your OKCupid profile and setting you up on dates). As another illustration, imagine telling an entrepreneur friend that superintelligences can kill us all, and then watching him take drastic actions that clearly indicate he's missing important nuances, all while he misunderstands and dismisses concerns you raise to him. Now reimagine these scenarios with your friends drastically more powerful than you.

This is my picture of what happens by default if we construct a recursively self-improving superintelligence by having it learn from human approval. The superintelligence would not be malevolent the way a paperclip maximizer would be, but for all intents and purposes might be.

14 comments

Comments sorted by top scores.

comment by Stuart_Armstrong · 2018-04-10T10:17:54.592Z · LW(p) · GW(p)

I think this is the wrong way of looking at it. Because in this analogy, the PA is "genuinely trying their hardest to optimize for your values", it's just poor at understanding these values. That problem is basically ignorance, and so by making the PA smarter or more aware, we can solve the problem.

But an AGI that fully understood your values, would still not optimise for them if it had a bad goal. The AGI is not well-intentioned-but-weird-in-implementation; its intentions themselves are alien/weird to us.

Replies from: zhukeepa
comment by zhukeepa · 2018-04-12T08:04:51.896Z · LW(p) · GW(p)

I wish I were clearer in my title that I'm not trying to reframe all misaligned AGI's, just a particular class of them. I agree that an AGI that fully understood your values would not optimize for them (and would not be "well-intentioned") if it had a bad goal.

That problem is basically ignorance, and so by making the PA smarter or more aware, we can solve the problem.

I think if we've correctly specified the values in an AGI, then I agree that when the AGI is smart enough it'll correctly optimize for our values. But it's not necessarily robust to scaling down, and I think it's likely to hit a weird place where it's trying and failing to optimize for our values. This post is about my intuitions for what that might look like.

Replies from: Stuart_Armstrong
comment by Stuart_Armstrong · 2018-04-12T10:13:36.218Z · LW(p) · GW(p)

Ok; within that subset of problems, I agree.

comment by Ben Pace (Benito) · 2018-04-09T23:58:03.281Z · LW(p) · GW(p)

I've curated this post for these reasons:

  • You put solid work into understanding central ideas in alignment, and published 4 posts simultaneously (1 [LW · GW], 2 [LW · GW], 3 [LW · GW]) communicating some key intuitions you've picked up. Doing this communicative work is really valuable and something I want to celebrate.
  • The posts helped me develop clear intuitions about corrigibility, approval-directed agents, and agent-foundations.
  • The posts are short, concrete, and easy to understand, which is the opposite of basically all the rest of the writing about AGI alignment.

My biggest hesitation(s) with curating this post:

  • I am somewhat hesitant to share simple intuition pumps about important topics, in case those intuition pumps are misleading.
  • I had a real hard time which of the four to pick, and have gone with the one that I thought people would find the clearest to read.

I was really excited that you wrote these posts, and learned a lot from them (plus the ensuing discussion in the comments).

Replies from: KatjaGrace, zhukeepa
comment by KatjaGrace · 2018-04-18T04:54:36.832Z · LW(p) · GW(p)
I am somewhat hesitant to share simple intuition pumps about important topics, in case those intuition pumps are misleading.

This sounds wrong to me. Do you expect considering such things freely to be misleading on net? I expect some intuition pumps to be misleading, but for considering all of the intuitions that we can find about a situation to be better than avoiding them.

Replies from: Benito, Raemon
comment by Ben Pace (Benito) · 2018-04-19T00:03:51.578Z · LW(p) · GW(p)

I feel like there are often big simplifications of complex ideas that just convey the wrong thing, and I was vaguely worried that in a field primarily dominated by things that are hard-to-read, things that are easy to understand will dominate the conversation even if they're pretty misguided. It's not a big worry for me here, but it was the biggest hesitation I had.

comment by Raemon · 2018-04-18T23:19:31.206Z · LW(p) · GW(p)

Not sure what Ben meant, but my own take is "sharing is fine, but intuition pumps without rigor backing them are not something we should curate regularly as an exemplar of what LW is trying to be"

comment by zhukeepa · 2018-04-10T02:55:59.733Z · LW(p) · GW(p)

Thanks a lot Ben! =D

I am somewhat hesitant to share simple intuition pumps about important topics, in case those intuition pumps are misleading.

On that note, Paul has recently written a blog post clarifying that his notion of "misaligned AI" does not coincide with what I wrote about here.

comment by John_Maxwell (John_Maxwell_IV) · 2018-04-04T00:33:38.376Z · LW(p) · GW(p)

In Superintelligence, Nick Bostrom talks about various "AI superpowers". One of these is "Social manipulation", which he summarizes as

Social and psychological modeling, manipulation, rhetoric persuasion

Strategic relevance:

  • Leverage external resources by recruiting human support
  • Enable a “boxed” AI to persuade its gatekeepers to let it out
  • Persuade states and organizations to adopt some course of action
  • AI can expropriate computational resources over the Internet

And Eliezer Yudkowsky writes:

There’s a popular concept of “intelligence” as book smarts, like calculus or chess, as opposed to say social skills. So people say that “it takes more than intelligence to succeed in human society”. But social skills reside in the brain, not the kidneys. When you think of intelligence, don’t think of a college professor, think of human beings; as opposed to chimpanzees. If you don’t have human intelligence, you’re not even in the game.

In order to have elite social skills, you need to be able to form accurate models about the thoughts & intentions of others. But being able to form accurate models about the thoughts & intentions of an overseer is exactly the ability we'd like to see in a corrigible AI.

If we can build AI systems that form those models without being goal-driven agents, maybe it's possible to have the benefits of elite social skills without the costs. I'm optimistic that this is the case--many of our most powerful model-building techniques don't really behave as though they have some kind of goal they are trying to achieve in the world.

comment by Ben Pace (Benito) · 2018-04-01T01:00:15.131Z · LW(p) · GW(p)

Minor point: You changed the title! The title used to be better.

Replies from: zhukeepa
comment by zhukeepa · 2018-04-01T01:22:46.933Z · LW(p) · GW(p)

How about now? :P

Replies from: Benito
comment by Ben Pace (Benito) · 2018-04-01T07:52:36.432Z · LW(p) · GW(p)

I've honestly forgotten the exact original wording, but I like this one more than the thing I complained about. (The post is super short and sweet and I liked having the title be a clear handle to the idea - an "AGI reframing" is not as good a pointer as "a well-intentioned non-neurotypical super-powerful assistant".)

comment by scarcegreengrass · 2018-04-10T18:19:31.013Z · LW(p) · GW(p)

I think this is a clever new way of phrasing the problem.

When you said 'friend that is more powerful than you', that also made me think of a parenting relationship. We can look at whether this well-intentioned personification of AGI would be a good parent to a human child. They might be able to give the child a lot of attention, a expensive education, and a lot of material resources, but they might take unorthodox actions in the course of pursuing human goals.

comment by Viktor Riabtsev (viktor-riabtsev-1) · 2018-04-01T21:31:41.048Z · LW(p) · GW(p)

So if not Human Approval, are we not saying it has to have certain levels of Self-Faith? I.e. it has to improve based off it own set of Ideals, that in the long run are what we actually want from a benevolent God?

I think this might be the Case, although of course, who am I to think something is unilaterally true...

So would not then, a set of Ideals need to be created which, for Us as Humans, makes us Better Humans; and based off that try to converge to a set of Ideals that an AGI can then inherit and start Self Improving on It's Own Ideas about Improvements.

We can't birth a Child, and then expect it to stay a Child. At some point, it must be Independent, Self-Sufficient and Sufficiently Rational.