Is friendly AI "trivial" if the AI cannot rewire human values?

post by Alerus · 2012-05-09T17:48:24.657Z · LW · GW · Legacy · 58 comments

I put "trivial" in quotes because there are obviously some exceptionally large technical achievements that would still need to occur to get here, but suppose we had an AI with a utilitarian utility function of maximizing subjective human well-being (meaning, well-being is not something as simple as physical sensation of "pleasure" and depends on the mental facts of each person) and let us also assume the AI can model this "well" (lets say at least as well as the best of us can deduce the values of another person for their well-being). Finally, we will also assume that the AI does not possess the ability to manually rewire the human brain to change what a human values. In other words, the ability for the AI to manipulate another person's values is limited by what we as humans are capable of today. Given all this, is there any concern we should have about making this AI; would it succeed in being a friendly AI?

One argument I can imagine for why this fails friendly AI is the AI would wire people up to virtual reality machines. However, I don't think that works very well, because a person (except Cypher from the Matrix) wouldn't appreciate being wired into a virtual reality machine and having their autonomy forcefully removed. This means the action does not succeed in maximizing their well-being.

But I am curious to hear what arguments exist for why such an AI might still fail as a friendly AI.

58 comments

Comments sorted by top scores.

comment by [deleted] · 2012-05-09T19:54:11.507Z · LW(p) · GW(p)

Finally, we will also assume that the AI does not possess the ability to manually rewire the human brain to change what a human values. In other words, the ability for the AI to manipulate another person's values is limited by what we as humans are capable of today. Given all this, is there any concern we should have about making this AI; would it succeed in being a friendly AI?

AAaiiiieee!

Can you please think about the emphasised phrase by yourself for 5 minutes (as measured by a physical clock)?

Replies from: faul_sname, Alerus
comment by faul_sname · 2012-05-10T02:35:51.763Z · LW(p) · GW(p)

That was approximately my reaction. Considering how easy it is to influence humans, it would be a nontrivial task to avoid unintentionally changing those values. I suspect that a significantly subhuman AI could change human values. One might argue that by increasing and redistributing wealth and improving medical care, our expert systems today are changing human value unintentionally. Google ads, which are automated and "learn" what the user likes, are specifically tailored to influence us, and do a remarkable job considering the relative simplicity of the algorithms used.

comment by Alerus · 2012-05-09T21:00:55.694Z · LW(p) · GW(p)

What is wrong with the statement? The idea I'm trying to portray is that I as a person now, cannot go and forcefully rewire another person's values. The only ability I have to try an affect them is to be persuasive in argument or perhaps being deceptive about certain things to try and get them to a different position (e.g., consider the state of politics).

In contrast, one of the concerns for the future is that an AI may have the technological ability to more directly manipulate a person. So the question I'm asking is: is the future technology at the disposal of an AI the only reason it could behave "badly?" under such a utility function?

Also, please avoid such comments. I am interested in having this discussion, but alluding to finding something wrong in what I have posted and not saying what you think it is, is profoundly unhelpful and useless to discussion.

Replies from: drethelin, DanArmak
comment by drethelin · 2012-05-09T21:23:15.116Z · LW(p) · GW(p)

Consider that humans have modified human values to results as different as nazism and as jainism.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:35:00.515Z · LW(p) · GW(p)

And yet humanity is resistant to large scale effects because we also combat changes in values that are destructive (like nazism). Are you suggesting that through persuasive means an AI could convert the values of all humanity to something unsavory? I think this is a bit too negative a view on humanity. You might suggest conditioning from birth, but this will result in outrage from the rest of humanity which the AI, by our utility definition, is trying to avoid.

Replies from: drethelin, None
comment by drethelin · 2012-05-09T21:36:53.527Z · LW(p) · GW(p)

If you don't think world war 2 was a large scale effect, then I don't know what to say to you.

Replies from: faul_sname, Alerus
comment by faul_sname · 2012-05-10T02:25:47.182Z · LW(p) · GW(p)

Compared to what the effects of total war between large powers just 20 years later would have been like, WWII was a relatively small-scale effect.

Replies from: JoshuaZ
comment by JoshuaZ · 2012-05-10T03:22:29.589Z · LW(p) · GW(p)

I don't think in context that's terribly relevant. The point that it was a large scale result. That it could end up having been much worse when technology got better doesn't really impact the point.

comment by Alerus · 2012-05-09T21:47:24.082Z · LW(p) · GW(p)

You make my point right there. World War 2. We went to war in defiance of nazis and refused to be assimilated. People in Germany didn't even like what the nazis were doing. And finally, the nazis didn't care about our outrage and death in the resulting war. An AI trying to maximize well-being, will care profoundly about that, by definition.

Replies from: shminux, JoshuaZ
comment by shminux · 2012-05-09T22:58:59.700Z · LW(p) · GW(p)

You seem to think that you are living in a magical fair universe. Just because nothing really really bad happened to you/us yet, doesn't mean it cannot.

Replies from: Alerus
comment by Alerus · 2012-05-09T23:37:58.882Z · LW(p) · GW(p)

I don't think I live in a fair universe at all. Regardless, acknowledging that we don't live in a fair universe doesn't support your claim that an AI would be able to radically change the values of all humans on earth without outrage from others through persuasion alone.

Replies from: faul_sname
comment by faul_sname · 2012-05-10T02:27:19.915Z · LW(p) · GW(p)

Humans can radically change the values of humans through weak social pressure alone.

Replies from: Alerus
comment by Alerus · 2012-05-10T02:42:18.497Z · LW(p) · GW(p)

I feel like I've already responded to this argument multiple times in various other responses I've made. If you think there's something I've overlooked in those responses let me know, but this seems like a restatement of things I've already addressed. Also, if there is something in one of the responses I've made with which you disagree and have a different reason than what's been presented, let me know.

comment by JoshuaZ · 2012-05-10T03:27:30.693Z · LW(p) · GW(p)

The point that other humans fought against it doesn't change the central point that a very large fraction of humans could have a radically different effective morality. Moreover, if Germany hadn't gone to war but had instead done the exact same thing to its internal minorities, most of the world likely would not have intervened.

If you don't like this example so much, one can just look at changing attitudes on many issues. See for example Pinker's book "The Better Angels of Our Nature" where he documents extreme changes in historical attitudes about the ethics of violence. For example, war is considered much more of a negative now than it was a few centuries ago. Going to war to gain territory is essentially unthinkable today. Similarly, attitudes about animals have changed a lot. In the Middle Ages, forms of entertainment that were considered normal included not just bear bating and similar actions but such crude behavior as lighting a cat on fire and seeing how long it took to die. Our moral attitudes are very much a product of our culture and how we are raised.

Replies from: Alerus
comment by Alerus · 2012-05-10T04:04:37.951Z · LW(p) · GW(p)

Most of our changes to where we are now seem to be a result of what works better in complex society and I therefore have difficulty accepting that a society in the highly advanced state it would be in by the time we had strong AI could be pushed to a non-productive doomsday set of values. So lets make the argument more clear then: what set of values do you think the AI could push us to through persuasion that would be effectively what we consider a doomsday scenario while and allowed the AI to more easily satisfy well-being?

Replies from: JoshuaZ
comment by JoshuaZ · 2012-05-10T04:39:28.989Z · LW(p) · GW(p)

I'm not sure why running a complex society needs to be a condition. If we all revert to hunter-gatherers then it still satisfies the essential conditions.

That's a problem even if it isn't a doomsday scenario. Changes in animal welfare attitudes would probably make most of us unhappy, but having a society where torturing cute animals to death wouldn't hurt running a complex society. Similarly, allowing infanticide would work fine (heck for that one I can think of some pretty decent arguments why we should allow it). And while not a doomsday scenario, other scenarios that could suck a lot can also be constructed. For example, you could have a situation where we're all stuck with 1950s gender roles. That would be really bad but wouldn't destroy a complex society.

Replies from: Alerus
comment by Alerus · 2012-05-10T13:47:27.550Z · LW(p) · GW(p)

Hunter gathers is not something sustainable for a large scale complex society. It is not a position we would favor at all and I'm struggling to see why an AI would try to make us value that set up or how you think a society with technology strong enough to make strong AI would be able to be convinced to it.

Views of killing animals is more flexible as the reason humans object to it seems to come from a level of innate compassion for life itself. So I could see that value being more manipulatable as a result. I don't see what that has to do with a doomsday set of values though.

1950s gener roles were abandoned because (1) women didn't like it (in which case maximizing people's well being would suggest not having such gender roles) and (2) it was less productive for society in that suppressing women limits the set of contributions to society.

I don't think you've presented here a set of doomsday values to which humans could be manipulated to holding by persuasion alone or demonstrated why they would be a set of values the AI would prefer humans to have for maximization.

comment by [deleted] · 2012-05-10T05:20:28.411Z · LW(p) · GW(p)

And yet humanity is resistant to large scale effects because we also combat changes in values that are destructive (like nazism).

Ah implicit belief in moral progress (or at least values and morality being preserved) and a universe where really bad things can't happen. I sometimes miss that.

The Nazis where defeated not because they where destructive but because the Soviet Union, UK and the US where stronger. Speaking of Stalin, how does Communism fit into your model?

Replies from: Alerus
comment by Alerus · 2012-05-10T13:58:16.137Z · LW(p) · GW(p)

Lets lose the silly straw man arguments. I've already explicitly commented on how I don't believe the universe is fair and I think from that it should be obvious that I don't think really bad things can't happen. As far as moral progress goes, I think it happens in so far as its functional. Morals that lead to more successful societies win the competition and stick around. This often happens to move societies (not necessarily all people in the society) toward greater tolerance of peoples and less violence because oppressing people and allowing for more violence tends to have bad effects internally in the society.

If we were weaker the Nazis could have won. That's not even the central point though. For kicks, lets assume the Nazis would have won the war. What does that mean though? It still means that other humans were is huge opposition and went to war over it causing enumerable deaths. After the nazis won, there would also surely be people wildly unhappy with the situation. This presents a serious problem for the AI trying to maximize well-being. It would not want to do things that led to mass outrage and opposition because that fails its own metrics.

Replies from: None
comment by [deleted] · 2012-05-10T15:24:36.379Z · LW(p) · GW(p)

After the nazis won, there would also surely be people wildly unhappy with the situation.

Consider what we think we know of the Nazis, are you sure about this one?

Replies from: Alerus
comment by Alerus · 2012-05-10T17:28:59.373Z · LW(p) · GW(p)

Considering there were many people in germany who vehemently disliked the nazis too (even ignoring jews), it seems like a pretty safe bet that after being conquered we wouldn't have suddenly viewed the nazis as great people. Why do you think otherwise?

Replies from: J_Taylor, None, Swimmy
comment by J_Taylor · 2012-05-10T17:49:49.616Z · LW(p) · GW(p)

The Japanese are rather fond of America, if I am not mistaken. I assume that it is not uncommon for the conquered to eventually grow satisfied with their conquerors.

Replies from: Alerus
comment by Alerus · 2012-05-10T19:56:00.567Z · LW(p) · GW(p)

We also didn't conquer Japan, we won the war. Those are two different things.

Replies from: J_Taylor
comment by J_Taylor · 2012-05-10T21:56:55.075Z · LW(p) · GW(p)

What sort of things would be different if it were the case that America conquered Japan?

Replies from: Alerus
comment by Alerus · 2012-05-10T22:16:06.705Z · LW(p) · GW(p)

Conquer is typically used to mean that you take over the government and run the country, not just win a war.

Replies from: J_Taylor, asr
comment by J_Taylor · 2012-05-10T22:24:11.359Z · LW(p) · GW(p)

Americans did rule Japan by military force for about five years after WWII ended, demilitarized the nation, and left behind a sympathetic government of American design. However, if you do not wish to use the word 'conquer' to describe such a process, that is your prerogative.

Replies from: Alerus
comment by Alerus · 2012-05-11T04:17:30.171Z · LW(p) · GW(p)

When you think of a nation conquering another, the US and Japan is really what comes to your mind? Are you honestly having trouble grasping the distinction I was making? Because personally, I'm really not interested in continuing an irrelevant semantics debate.

comment by asr · 2012-05-10T22:17:50.785Z · LW(p) · GW(p)

The US ran the Japanese government for a period of several years. I think you mean to add something about "run the country without intent to transfer power back to the locals".

Replies from: Alerus
comment by Alerus · 2012-05-11T04:07:17.108Z · LW(p) · GW(p)

Yes. I find it odd that this argument is derailed into demanding a discussion on the finer points of the semantics for "conquer."

Replies from: asr
comment by asr · 2012-05-11T07:31:28.817Z · LW(p) · GW(p)

A confession: Often when reading LW, I will notice some claim that seems wrong, and will respond, without reading the thread context carefully or checking back far enough to understand exactly why a given point came up. This results in an inadvertent tendency to nit-pick, for which I apologize.

Replies from: Alerus
comment by Alerus · 2012-05-11T15:22:42.292Z · LW(p) · GW(p)

I appreciate that sentiment and I'll also add that I appreciate that even in your prior post you made an effort to suggest what you thought I was driving at.

comment by [deleted] · 2012-05-10T17:50:53.459Z · LW(p) · GW(p)

For the Poles at least I fear it probable not many would be around say 20 years after victory.

Replies from: Alerus
comment by Alerus · 2012-05-10T20:06:55.492Z · LW(p) · GW(p)

And as for the others? Or are you saying the AI trying to maximize well-being will try and succeed in effectively wiping out everyone and then condition future generations to have the desired easily maximized values? If so, this behavior is conditioned on the idea that the AI could be very confident in its ability to do so, because otherwise the chance of failing and the cost of war in expected value of human well-being would massively drop the expected value. I think you should also make clear what you think these values might end up being to which it will try to change human values to more easily maximize.

comment by Swimmy · 2012-05-10T18:15:28.438Z · LW(p) · GW(p)

It's irrelevant. In a world of world-destroying technologies, a really bad thing happening for only a small amount of time is all it takes. The Cold War wasn't even close to the horror of Nazi domination (probably)--there were still lots of happy people with decent governments in the west! But everyone still could have died.

What if Nazis had developed nuclear weapons? What if the AI self-reproduces, without self-improving, such that the Big Bad they're supporting has an army of super-efficient researchers and engineers? What if they had gotten to the hydrogen bomb around the same time the US had gotten the atom bomb? What if the Big Bad develops nanomachines, programmable to self-replicate and destroy anyone who opposes, or who passes a certain boundary? What kind of rebellion or assassination attempt could stand up to that? What if the humans want the AI, rather than another human, to be the leader of their Big Bad Movement, making their evil leader both easily replicable and immune to nanomachine destruction?

Hell, what if the AI gets no more competent or powerful than a human? It can still, in the right position, start a thermonuclear war, just the same as high-level weapons techs or--hell!--technical errors can. Talented spies can make it to sufficiently high levels of government operation; why couldn't a machine do so? Or hire a spy to do so?

And if the machine thinks that's the best way to make people happy (for whatever horrible reason--perhaps it is convinced by the Repugnant Conclusion and wants to maximize utility by wiping out all the immiserated Russians), we're still in trouble.

However, if you're trying to describe an AI that is set to maximize human value, understands the complexities of the human mind, and won't make such mistakes, then you are describing friendly AI.

Edit: In other words, I contend that the future threat of General AI is not in modifying humans with nanotechnology. It is in simple general ability to shape the world, even if that only means manipulating objects using current technologies. If we're defining "intelligence" as the ability to manipulate atoms to shape the world according to our bliss points, a machine that can think thousands of times faster than humans will be able to do so at least hundreds of times better than humans. This is especially true if it can replicate, which, given this hypothesis, it will almost certainly be able to. If we add intelligence explosion to the mix, we're in big trouble.

Replies from: Alerus
comment by Alerus · 2012-05-10T20:22:24.877Z · LW(p) · GW(p)

You're missing the point of talking about opposition. The AI doesn't want the outcome of opposition because that has terrible effects on the well-being its trying to maximize, unlike the nazis. This isn't about winning the war, its about the consequence of war on the measured well-being of people and other people who live in a society where an AI would kill people for what amounted to thought-crime.

And if the machine thinks that's the best way to make people happy (for whatever horrible reason--perhaps it is convinced by the Repugnant Conclusion and wants to maximize utility by wiping out all the immiserated Russians), we're still in trouble.

This specifically violates the assumption that the AI has well modeled how any given human measures their well-being.

However, if you're trying to describe an AI that is set to maximize human value, understands the complexities of the human mind, and won't make such mistakes, then you are describing friendly AI.

It is the assumption that it models human well-being at least as well as the best a human can model the well-being function of another. However, this constraint by itself does not solve friendly AI, because in a less constrained problem than the one I outlined, the most common response for an AI trying to maximize what humans value is that it will change and rewire what humans value to something more easy to maximize. The entire purpose of this post is to question whether it could achieve this without the ability to manually rewire human values (e.g., could this be done through persuasion?). In other words, you're claiming friendly AI is solved more easily than the constrained question I posed in the post.

Replies from: Swimmy
comment by Swimmy · 2012-05-11T02:01:57.592Z · LW(p) · GW(p)

Are you trying to argue that, of all the humans who have done horrible horrible things, not a single one of them 1) modeled other humans at the average or above-average level that humans usually model each other, and 2) not a single one of them thought they were trying to make the world better off? Or are you trying to argue that not a single one of them ever caused an existential threat?

My guess is that Lenin, for instance, had an above-average human-modeling mind and thought he was taking the first steps of bringing the whole world into a new prosperous era free from class war and imperialism. And he was wrong and thousands of people died. The kulaks opposed, in the form of destroying their farms. Lenin probably didn't "want the outcome of opposition," but that didn't stop him from thinking mass slaughter was the solution.

The ability to model the well-being of humans and the "friendliness" of the AI are the same thing, provided the AI is programmed to maximize that well-being value. If your AI can't ever make mistakes like that, it's a friendly AI. If it can, it's trouble whether or not it can alter human values.

comment by DanArmak · 2012-05-09T21:36:15.497Z · LW(p) · GW(p)

Consider that every human who ever existed, was shaped purely by environment + genes.

Consider how much humans have achieved merely by controlling the environment: converting people to insane religions which they are willing to die and kill for, making torturers, "the banality of evil", etc. etc.

Now imagine what an entity could achieve with that plus 1) complete understanding of how the brain is shaped by the environment and/or 2) complete control of the environment (via VR, smart dust, whatever) for a human from age 0 onwards.

I think the conservative assumption is that any mind we would recognize as human, and many we wouldn't, could be produced by such an optimization process. You're not limiting your AI at all.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:53:41.163Z · LW(p) · GW(p)

But the AI isn't being dropped into a completely undeveloped society. It will be dropped into an extremely developed society with values already existing. If the AI were dropped back into the era of early man, I could see major concern. I don't see humanity having the values we've developed being radically and entirely changed into something we consider so unsavory by persuasion alone. That doesn't mean no one could be affected, but I can't see such a thing going down without outrage from large sects of humanity; which is not what the AI wants.

Replies from: DanArmak
comment by DanArmak · 2012-05-09T22:25:17.801Z · LW(p) · GW(p)

You underestimate "persuasion alone". Please consider that (by your definition) all human opinions on all subjects that have existed to date, have been created pretty much "by persuasion alone".

Also, I don't want to live in a world where what I'm allowed to do or be is constrained by whether it provokes "outrages from large sects of humanity". There are plenty of sects (properly so called ;-) today that don't want me to continue existing even the way I already am, at least not without major brainwashing.

Replies from: Alerus
comment by Alerus · 2012-05-09T22:42:21.395Z · LW(p) · GW(p)

All human opinions cannot be created by persuasion alone because opinions have to start somewhere. People can and do think for themselves and that's what creates opinions. Then they might persuade people to have these opinions as well, but clearly persuasion is not the sole source and even then it's not like persuasion is a one-way process where you hit the persuade button and the other person is switched. It seems that your argument is that any human can be persuaded to any opinion at any time and I just can't buy that. Humans are malleable and we've made a huge number of mistakes in the past, but I don't see us as so bad that anyone can have their mind changed to anything regardless of the merit behind it. This entire site is based around getting people to not be arbitrarily malleable and to require rationality in making decisions—that there are objective conclusions and we should strive for them. Is this site and community a failure then? Are all of the people subject to mere persuasion in spite of rationality and cannot think for themselves?

Regarding actions that cause outrage I never said you were constrained by the outrage of others. I said an AI that maximizes human well-being is not going to take actions that cause extreme outrage.

Replies from: DanArmak
comment by DanArmak · 2012-05-10T00:08:00.184Z · LW(p) · GW(p)

All human opinions cannot be created by persuasion alone because opinions have to start somewhere. People can and do think for themselves and that's what creates opinions.

This is completely wrong. Again, you give "persuasion" a very narrow scope.

A baby is born without language, certainly without many opinions. It can be shaped by its environment ("persuasion") to be almost anything. Certainly, very few of the extremely diverse cultures and sub-cultures known from history have had any trouble raising their kids to behave like the adults, with only a small typical proportion of adolescents who left for another society. And these people had no understanding of how the brain really works - unlike what a superintelligent AI might have.

Short version: it doesn't matter if people do think for themselves, because they only get to think about their sensory inputs and the AI can control those. Even a perfect Bayesian superintelligence would reach any conclusion you wished if you truly fully controlled all the information it ever received (as long as it had no priors of 0 or 1).

This entire site is based around getting people to not be arbitrarily malleable and to require rationality in making decisions [...] Is this site and community a failure then?

If you end up in an environment controlled by an unfriendly AI, having read this site won't help you; it's game over. LW rationality skills work in some worlds, not in any possible world.

Regarding actions that cause outrage I never said you were constrained by the outrage of others. I said an AI that maximizes human well-being is not going to take actions that cause extreme outrage.

How is this different from saying it's not going to let me take actions that cause extreme outrage? I hope you aren't planning on building an AI that has a sense of personal responsibility and doesn't care if humans subvert its utility function as long as it didn't cause them to do so.

Replies from: Alerus
comment by Alerus · 2012-05-10T02:02:43.351Z · LW(p) · GW(p)

There is a profound difference between being persuasive and manipulating all sensory input of a human. Is your argument not that it would try to persuade but that an AI would hook up all humans to a computer that controlled everything we perceived? If you want to make that your argument, I'm game for discussing it, but I think it should be made clear that this is a very different argument than an AI trying to change people's minds through persuasion. But lets discuss it. This suggestion of manipulating the senses of humans seems to imply a massive use of technology and integration of the technology by the AI not available today, but that's okay, we should expect technology to improve incredibly by the time we can make strong AI. But so long as we're assuming that such huge amounts of improved technology with large integration is available and would allow the AI to pull the wool over everyone's eyes, we must also consider that humans have made use of that technology themselves to better themselves and provide wildly intelligent computer security systems such that it seems a stretch to me to posit that an AI could do this without anyone noticing.

How is this different from saying it's not going to let me take actions that cause extreme outrage? I hope you aren't planning on building an AI that has a sense of personal responsibility and doesn't care if humans subvert its utility function as long as it didn't cause them to do so.

I suppose if your actions were extreme enough in the outrage they caused we might make a case for those actions needing to be thwarted, even by the reasoning of the AI. I don't know you, but my guess is you're thinking perhaps of religious fundamentalists feelings about you? Such outrage on its own is (1) somewhat limited and counterbalanced by others and (2) counter productive for humanity to act upon in which case the better argument is not to thwart your actions but work toward behavior of tolerance. But lets contrast this to an AI trying to effectively replace mankind with easily satisfied humans and consider how people would respond to that. I think its clear that humans would work toward shutting such an AI down and would respond with extreme concern for their livelihood. The fact that we're sitting her talking about how this is doomsday scenario seems to be evidence of that concern. Given that, it just doesn't seem to be in the AIs interest to make that choice; it would simply cause too much of a collapse in the well-being of humanity with their profound concern for the situation.

comment by Normal_Anomaly · 2012-05-09T21:20:33.163Z · LW(p) · GW(p)

suppose we had an AI with a utilitarian utility function of maximizing subjective human well-being (meaning, well-being is not something as simple as physical sensation of "pleasure" and depends on the mental facts of each person) and let us also assume the AI can model this "well" (lets say at least as well as the best of us can deduce the values of another person for their well-being)

You've crammed all the difficulty of FAI into this sentence. An additional limit on how much it can manipulate us does little if anything to make this part easier, and adds the additional complication of how strict this limitation should be. The question of how much FAI would manipulate us is an interesting one, but either it's a small part of the problem or it's something that will be subsumed in the main question of "what do we want?". By the latter I mean that we may decide that the best way to decide how much FAI should change our values is to have it calculate our CEV, the same way that the FAI will decide what economic system to implement.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:39:11.477Z · LW(p) · GW(p)

This is not meant to be a resolution to FAI since you can't stop technology. It's meant to highlight whether the bad behavior of AI ends up being due to future technology to more directly change humanity. I'm asking the question because the answer to this may give insights as to how to tackle the problem.

comment by DanielLC · 2012-05-09T20:11:25.719Z · LW(p) · GW(p)

I'd suggest reading Failed Utopia #4-2.

One problem is that if it can create new people, any rules about changing people would be pointless. If it cannot create new people, then it ends up with a Utopia for 6 billion people, which is nothing compared to what could have been.

This could be fixed by letting it rewire human brains, but limiting it to doing what humans would be okay with, if it didn't rewire their brains. This is better, but it still runs into problems in that people wouldn't fully understand what's going on. What you need to do is program it so that it does what people would like if they were smarter, faster, and more the people they wish they were. In other words, use CEV.

Also, it's very hard to define what exactly constitutes "rewiring a human brain". If you make it too general, the AI can't do anything, because that would affect human brains. If you make it too specific, the AI would have some slight limitations on how exactly it messes with people's minds.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:18:58.685Z · LW(p) · GW(p)

Thanks for the link, I'll give it a read.

Creating new people is potentially a problem, but I'm not entirely convinced. Let me elaborate. When you say:

What you need to do is program it so that it does what people would like if they were smarter, faster, and more the people they wish they were. In other words, use CEV.

Doesn't this kind of restate in different words that it models human well-being and tries to maximize that? I imagine when you phrased it this way that such an AI wouldn't create new people that are easier to maximize because that isn't what humans would want. And if that's not what humans would want doesn't that just mean it's negatively viewed in their well-being and my original definition suffices? Assuming humans don't want the AI to make new people that are simply easier to maximize, if it created a new person, all people on the earth view this negatively and their well-being drops. In fact, it may lead to humans shutting the AI down, so the AI deduces that it cannot create new people that are easier to maximize. The only possible hole in that I see is if the AI could suddenly create an enormous number of people at once..

Also, it's very hard to define what exactly constitutes "rewiring a human brain". If you make it too general, the AI can't do anything, because that would affect human brains. If you make it too specific, the AI would have some slight limitations on how exactly it messes with people's minds.

Indeed it's difficult to say precisely, that's why I used what we can do now as analogy. I can't really rewire a person's values at all except through persuasion or other such methods. Even our best neuroscientists can't do that unless I'm ignorant to some profound advances. The most we can really do is tweak pleasure centers (which as I stated isn't the metric for well-being) or effectively break the brain so the person is non-operational, but I'd argue that non-operational humans have effectively zero measure of well-being anyway (for similar reasons as to why I'd say a bug has a lower scale of well-being than a human does).

Replies from: DanielLC
comment by DanielLC · 2012-05-09T23:09:42.270Z · LW(p) · GW(p)

Assuming humans don't want the AI to make new people that are simply easier to maximize, if it created a new person, all people on the earth view this negatively and their well-being drops.

I'm not sure how common it is, but I at least consider total well-being to be important. The more people the better. The easier to make these people happy, the better.

Indeed it's difficult to say precisely, that's why I used what we can do now as analogy. I can't really rewire a person's values at all except through persuasion or other such methods.

An AI is much better at persuasion than you are. It would pretty much be able to convince you whatever it wants.

Even our best neuroscientists can't do that unless I'm ignorant to some profound advances.

Our best neuroscientists are still mere mortals. Also, even among mere mortals, making small changes towards someones values are not difficult, and I don't think significant changes are impossible. For example, the consumer diamond industry would be virtually non-existant if De Beers didn't convince people to want diamonds.

Replies from: ArisKatsaris, Alerus
comment by ArisKatsaris · 2012-05-10T01:16:28.005Z · LW(p) · GW(p)

The more people the better.

The more people in what? Any particular moment in time? The complete timeline of any given Everett Branch? The whole multiverse?

Between an Everett branch of 10 billion people, and ten Everett branches of 1 billion people each, which do you prefer?

Between 10 billion people that live in the same century, and one billion people per century over a span of ten centuries, which do you prefer?

Replies from: DanielLC
comment by DanielLC · 2012-05-10T02:15:06.415Z · LW(p) · GW(p)

The whole multiverse.

comment by Alerus · 2012-05-09T23:48:01.063Z · LW(p) · GW(p)

I'm not sure how common it is, but I at least consider total well-being to be important. The more people the better. The easier to make these people happy, the better.

You must also consider that well-being need not be defined as a positive function. Even if it wasn't, if the gain of adding a person was less than drop in well-being of others, it wouldn't be beneficial unless the AI was able to without prevention, create many more such people.

An AI is much better at persuasion than you are. It would pretty much be able to convince you whatever it wants.

I'm sure it'd be better than me (unless I'm also heavily augmented by technology, but we can avoid that issue for now). On what grounds can you say that it'd be able to persuade me to anything it wants? Intelligence doesn't mean you can do anything and think this needs to be justified.

Our best neuroscientists are still mere mortals. Also, even among mere mortals, making small changes towards someones values are not difficult, and I don't think significant changes are impossible. For example, the consumer diamond industry would be virtually non-existant if De Beers didn't convince people to want diamonds.

I know they're mere mortals. We're operating under the assumption that the AI's methods of value manipulation are limited to what we can do ourselves, in which case rewiring is not something we can do with any great affect. The point of the assumption is to ask what the AI could do without more direct manipulation. To that end, only persuasion has been offered and as I've stated, I'm not seeing a compelling argument for why an AI could persuade anyone to anything.

Replies from: DanielLC
comment by DanielLC · 2012-05-10T00:26:10.691Z · LW(p) · GW(p)

Even if it wasn't, if the gain of adding a person was less than drop in well-being of others, it wouldn't be beneficial unless the AI was able to without prevention, create many more such people.

Do you honestly think a universe the size of ours can only support six billion people before reaching the point of diminishing returns?

We're operating under the assumption that the AI's methods of value manipulation are limited to what we can do ourselves, in which case rewiring is not something we can do with any great affect.

If you allow it to use the same tools but better, it will be enough. If you don't, it's likely to only try to do things humans would do, on the basis that they're not smart enough to do what they really want done.

Replies from: Alerus
comment by Alerus · 2012-05-10T01:29:23.253Z · LW(p) · GW(p)

Do you honestly think a universe the size of ours can only support six billion people before reaching the point of diminishing returns?

That's not my point. The point is people aren't going to be happy if an AI starts making people that are easier to maximize for the sole reason that they're easier to maximize. This will suggest a problem to us by the very virtue that we are discussing hypotheticals where doing so is considered a problem by us.

If you allow it to use the same tools but better, it will be enough. If you don't, it's likely to only try to do things humans would do, on the basis that they're not smart enough to do what they really want done.

You seem to be trying to break the hypothetical assumption on the basis that I have not specified a complete criteria that would prevent an AI from rewiring the human brain. I'm not interested in trying to find a set of rules that would prevent an AI from rewiring human's brain (and I never tried to provide any, that's why it's called an assumption), because I'm not posing that as a solution to the problem. I've made this assumption to try and generate discussion all the problems where it will break down since typically discussion seems to stop at "it will rewire us". Trying to assert "yeah but it would rewire because you haven't strongly specified how it couldn't" really isn't relevant to what I'm asking since I'm trying to get specifically at what it could do besides that.

comment by ErikM · 2012-05-09T19:57:38.316Z · LW(p) · GW(p)

"Finally, we will also assume that the AI does not possess the ability to manually rewire the human brain to change what a human values. In other words, the ability for the AI to manipulate another person's values is limited by what we as humans are capable of today."

I argue that we as humans are capable of a lot of that, and the AI may be able to think faster and draw upon a larger store of knowledge of human interaction.

Furthermore, what justifies this assumption? If we assume a limit that the AI won't manipulate me any more than Bob across the street will manipulate me, then yes the AI is safe, but that limit seems very theoretical. A higher limit that the AI won't manipulate me more than than the most manipulative person in the world isn't very reassuring, either.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:23:11.785Z · LW(p) · GW(p)

Can you give examples of what you think humans capability to rewire another's values are?

As for what justifies the assumption? Nothing. I'm not asking it specifically because I don't think AIs will have it, I'm asking it so we can identify where the real problem lies. That is, I'm curious whether the real problem in terms of AI behavior being bad is entirely specific to advances in biological technology to which eventual AIs will have access, but we don't today. If we can conclude this is the case, it might help us in understanding how to tackle the problem. Another way to think of the question I'm asking is take such an AI robot and drop it into todays society. Will it start behaving badly immediately, or will it have to develop technology we don't have today before it can behave badly?

Replies from: shminux
comment by shminux · 2012-05-09T21:33:16.156Z · LW(p) · GW(p)

Can you give examples of what you think humans capability to rewire another's values are?

As plenty of religious figures have shown over the years, this capability is virtually unlimited. An AI would just have to start a new religion, or take over an existing one and adapt it to its liking.

Replies from: Alerus
comment by Alerus · 2012-05-09T21:43:53.594Z · LW(p) · GW(p)

And yet as time goes on civilization is progressing to more secular values. It will be interesting to see where we are by the time strong AI is possible, especially since we undoubtedly will be changing ourselves to improve our own capabilities. As I said in one of the other comments, I think assuming that humanity can in totality be persuaded to unsavory values, even through religion, is too negative a view of humanity. Humanity's history with religion is also filled with defiance and an AI that values human well-being will not be pleased with the outrage reaction as it tries to gain followers through persuasive means.

comment by D_Alex · 2012-05-11T06:31:18.637Z · LW(p) · GW(p)

I am saddened by the amount of downvotes on Alerus's well written and provocative posts. He made a positive contribution to the discussion, and should not be discouraged, IMO.

Replies from: Alerus
comment by Alerus · 2012-05-11T16:09:01.673Z · LW(p) · GW(p)

Thanks, I appreciate that. I have no problem with people disagreeing with me as confronting disagreement is how people (self included) grow. However, I was taken aback by the amount of down voting I received merely for disagreeing with people here and the fact that by merely choosing to respond to people's arguments it would effectively guarantee even more down votes—a system tied to how much you can participate in the community—made it more concerning to me. At least on the discussion board side of the site, I expected down voting to be reserved for posts that were derailing topics, flaming, ignoring arguments presented to them, etc., not for posts with which one disagreed. As someone who does academic research in AI, I thought this could be a fun lively online community to discuss that, but having my discussion board topic posting privileges removed because people did not agree with things I said (and the main post didn't even assert anything, it asked for feedback), I've reconsidered that. I'm glad to see not all people here think this was an appropriate use of down voting, but I feel like the community at large has spoken with regards to how they use that and when this thread ends I'll probably be moving on.

Thanks for you support though, I do appreciate that.