Do what we mean vs. do what we say

post by rohinmshah · 2018-08-30T22:03:27.665Z · score: 30 (15 votes) · LW · GW · 12 comments

Written quickly after a CHAI meeting on the topic, haven't thought through it in depth.

If we write down an explicit utility function and have an AI optimize that, we expect that a superintelligent AI would end up doing something catastrophic, not because it misunderstands what humans want, but because it doesn't care -- it is trying to optimize the function that was written down. It is doing what we said instead of doing what we meant.

An approach like Inverse Reward Design instead says that we should take the human's written down utility function as an observation about the true reward function, and infer a distribution over true reward functions. This agent is "doing what we mean" instead of doing what we said.

This suggests a potential definition -- in a "do what we mean" system, the thing that is being optimized is a latent variable, whereas in a "do what we say" system, it is explicitly specified. Note that "latent" need not mean that you have a probability distribution over it, it just needs to be hidden information. For example, if I had to categorize iterated distillation and amplification, it would be as a "do what we mean" system where the thing being optimized is implicit in the policy of the human and is never made fully certain.

However, this doesn't imply that we want to build a system that exclusively does what we mean. For example, with IRD, if the true reward function is not in the space of reward functions that we consider (perhaps because it depends on a feature that we didn't have), you can get arbitrarily bad outcomes (see the problem of fully updated deference). One idea would be to have a "do what we mean" core, which we expect will usually do good things, but have a "do what we say" subsystem that adds an extra layer of safety. For example, even if the "do what we mean" part is completely sure about the human utility function and knows we are making a mistake, the AI will still shut down if we ask it to because of the "do what we say" part. This seems to be the idea in MIRI's version of corrigibility.

I'd be interested to see disagreements with the definition of "do what we mean" as optimizing a latent variable. I'd also be interested to hear how "corrigibility" and "alignment" relate to these concepts, if at all. For example, it seems like MIRI's corrigibility is closer to "do what we say" while Paul's corrigibility is closer to "do what we mean".

12 comments

Comments sorted by top scores.

comment by Dagon · 2018-08-31T18:24:26.729Z · score: 4 (3 votes) · LW(p) · GW(p)

There's another layer of uncertainty here. For sufficiently general and long-term preferences, it's not clear that "do what we mean" is sufficient either. None of us knows what we want, so we what we mean isn't even very good evidence of what we want.

"do what I would want to mean" is closer, but figuring out the counterfactuals for "would" that preserve "I" is not easy.

comment by Charlie Steiner · 2018-08-31T21:04:22.321Z · score: 1 (1 votes) · LW(p) · GW(p)

Agreed. Humans don't really have utility functions. We might try to get around this by having the AI learn how humans would like to be interpreted as having a utility function, and how they would like that to be interpreted, and so on in an infinite tower of reflection, but that doesn't seem very practical or desirable.

I think there was an old Wei Dai post on "artificial philosophy" that was about this problem? The idea is we want the AI to collapse this infinite tower by learning the philosophical considerations that generate it, then use that knowledge to learn its preferences from humans.

comment by rohinmshah · 2018-09-01T18:31:57.915Z · score: 1 (1 votes) · LW(p) · GW(p)

Just don't ask your AI system to optimize for general and long-term preferences without a way for you to say "actually, stop, I changed my mind".

Like, if someone tells me that they want me to protect nature, I know that in effect they mean "Take actions to protect nature right now, but don't do anything super drastic that would conflict with other things I care about, and if I change my mind in the future, defer to that change, etc." I think a good "do what you mean" system would capture all of that. This isn't implied by my definition of course, but I think that a system where the specification is latent and uncertain could have this property.

comment by Nebu · 2018-09-10T05:43:30.900Z · score: 1 (1 votes) · LW(p) · GW(p)
Just don't ask your AI system to optimize for general and long-term preferences without a way for you to say "actually, stop, I changed my mind".

I believe that reduces to "solve the Friendly AI problem".

comment by rohinmshah · 2018-09-10T23:44:36.952Z · score: 3 (2 votes) · LW(p) · GW(p)

(Pedantic note: the right way to say that is "the Friendly AI problem reduces to that".)

I'm replying to the quote from the first comment:

For sufficiently general and long-term preferences, it's not clear that "do what we mean" is sufficient either. None of us knows what we want, so we what we mean isn't even very good evidence of what we want.

What I'm trying to say is that once you have a "do what we mean" system, then don't explicitly ask your AI system to optimize for general and long-term preferences without a way for you to say "actually, stop, I changed my mind".

I claim that the hard part there is in building a "do what we mean" system, not in the "don't explicitly ask for a bad thing" part.

comment by TurnTrout · 2018-08-31T03:57:00.204Z · score: 3 (2 votes) · LW(p) · GW(p)

Perhaps "do what we say" is more like "know when the outside view says you've incorrectly converged to the wrong value function, so we're probably right and you should listen to us".

comment by Charlie Steiner · 2018-08-31T20:44:30.939Z · score: 3 (2 votes) · LW(p) · GW(p)

It's somewhat more subtle than that. The ideal (and maybe impossible) corrigible AI should protect us even if we accidentally give the AI the wrong process for figuring out what to value. It should protect us even if the AI becomes omniscient.

If the AI knows vastly more than we do, there's no sense in which we are providing extra evidence or an information-carrying "outside view". We are instead just registering a sort of complaint and hoping we've programmed the AI to listen.

I'm still not convinced that such a sort of corrigibility is in any way distinct from some extra complications in the process we give the AI for figuring out what to value.

comment by TurnTrout · 2018-09-01T02:44:28.734Z · score: 4 (2 votes) · LW(p) · GW(p)

The outside view I had in mind wasn't with respect to its knowledge, but to empirical data on how often its exact value-learning algorithm converges to the correct set of preferences for agents-like-us. That feels different.

comment by rohinmshah · 2018-09-01T18:34:05.248Z · score: 1 (1 votes) · LW(p) · GW(p)

I assume you're talking about the particular "do what we say" subsystem described in the second last paragraph? If so, that seems plausibly right.

comment by Pattern · 2018-08-30T23:12:49.606Z · score: 2 (1 votes) · LW(p) · GW(p)

What we say: "Follow the recipe"

What we mean: "Make tasty, edible, food, with the ingredients provided, after verifying they are what they're supposed to be, etc. "

I think this [LW · GW] is related, although it's about getting the AI to ask humans questions about what to value.

comment by avturchin · 2018-09-01T09:20:14.751Z · score: 1 (1 votes) · LW(p) · GW(p)

This approach ignores choice. To have an utility function is not enough to make a choice, and what I say is an act of making a choice.

For example, I have hidden value function (apples = 0.5 and oranges =0.5). I ask my home robot to bring me an apple. In that moment I made a choice between equally preferable preferences.

But my home robot would ignore my choice and bring me half of apple and half of orange, because this was my value function before making the choice.

In that case, I will be not satisfied as I will feel that robot ignores my moral efforts of making a choice, and I value my choices. Also, after making the choice my preferences will be updated, so the robot should decide which my utility function should be used: before the choice or after.

comment by Dagon · 2018-09-01T15:16:32.725Z · score: 2 (1 votes) · LW(p) · GW(p)

(I don't think humans have consistent utility functions; we're broken that way. If we did...)

The robot should know your utility function(s) well enough to know that you'd choose apple this time, and orange at some future time.