Posts

Comments

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-25T07:28:31.115Z · LW · GW

[Policymakers]

"If we imagine a space in which all possible minds can be represented, we must imagine all human minds as constituting a small and fairly tight cluster within that space. The personality differences between Hannah Arendt and Benny Hill might seem vast to us, but this is because the scale bar in our intuitive judgment is calibrated on the existing human distribution. In the wider space of all logical possibilities, these two personalities are close neighbors. In terms of neural architecture, at least, Ms. Arendt and Mr. Hill are nearly identical. Imagine their brains laying side by side in quiet repose. The differences would appear minor and you would quite readily recognize them as two of a kind; you might even be unable to tell which brain was whose.

There is a common tendency to anthropomorphize the motivations of intelligent systems in which there is really no ground for expecting human-like drives and passions (“My car really didn’t want to start this morning”). Eliezer Yudkowsky gives a nice illustration of this phenomenon:

Back in the era of pulp science fiction, magazine covers occasionally depicted a sentient monstrous alien—colloquially known as a bug-eyed monster (BEM)—carrying off an attractive human female in a torn dress. It would seem the artist believed that a nonhumanoid alien, with a wholly different evolutionary history, would sexually desire human females … Probably the artist did not ask whether a giant bug perceives human females as attractive. Rather, a human female in a torn dress is sexy—inherently so, as an intrinsic property. They who made this mistake did not think about the insectoid’s mind: they focused on the woman’s torn dress. If the dress were not torn, the woman would be less sexy; the BEM does not enter into it. (Yudkowsky 2008)

An artificial intelligence can be far less human-like in its motivations than a space alien. The extraterrestrial (let us assume) is a biological creature who has arisen through a process of evolution and may therefore be expected to have the kinds of motivation typical of evolved creatures. For example, it would not be hugely surprising to find that some random intelligent alien would have motives related to the attaining or avoiding of food, air, temperature, energy expenditure, the threat or occurrence of bodily injury, disease, predators, reproduction, or protection of offspring. A member of an intelligent social species might also have motivations related to cooperation and competition: like us, it might show in-group loyalty, a resentment of free-riders, perhaps even a concern with reputation and appearance.

By contrast, an artificial mind need not care intrinsically about any of those things, not even to the slightest degree. One can easily conceive of an artificial intelligence whose sole fundamental goal is to count the grains of sand on Boracay, or to calculate decimal places of pi indefinitely, or to maximize the total number of paperclips in its future lightcone. In fact, it would be easier to create an AI with simple goals like these, than to build one that has a humanlike set of values and dispositions."

[Taken from Nick Bostrom's 2012 paper, The Superintelligent Will]

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-14T17:12:51.639Z · LW · GW

[Policy makers & ML researchers]

Expecting AI to automatically care about humanity is like expecting a man to automatically care about a rock. Just as the man only cares about the rock insofar as it can help him achieve his goals, the AI only cares about humanity insofar as it can help it achieve its goals. If we want an AI to care about humanity, we must program it to do so. AI safety is about making sure we get this programming right. We may only get one chance.
 

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-14T16:49:20.635Z · LW · GW

[Policy makers & ML researchers]

Our goal is human flourishing. AI’s job is to stop at nothing to accomplish its understanding of our goal. AI safety is about making sure we’re really good at explaining ourselves.

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-14T16:06:05.877Z · LW · GW

[Policy makers & ML researchers]

AI safety is about developing an AI that understands not what we say, but what we mean. And it’s about doing so without relying on the things that we take for granted in inter-human communication: shared evolutionary history, shared experiences, and shared values. If we fail, a powerful AI could decide to maximize the number of people that see an ad by ensuring that ad is all that people see. AI could decide to reduce deaths by reducing births. AI could decide to end world hunger by ending the world. 

(The first line is a slightly tweaked version of a different post by Linda Linsefors, so credit to her for that part.)

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-13T05:11:20.370Z · LW · GW

Thanks Trevor - appreciate the support! Right back at you.

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-05-12T13:12:18.042Z · LW · GW

[Policy makers & ML researchers]

"There isn’t any spark of compassion that automatically imbues computers with respect for other sentients once they cross a certain capability threshold. If you want compassion, you have to program it in" (Nate Soares). Given that we can't agree on whether a straw has two holes or one...We should probably start thinking about how program compassion into a computer.

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-28T05:56:20.146Z · LW · GW

[Policy makers & ML researchers]

Expecting AI to know what is best for humans is like expecting your microwave to know how to ride a bike.

[Insert call to action]

-

Expecting AI to want what is best for humans is like expecting your calculator to have a preference for jazz.

[Insert call to action]

(I could imagine a series riffing based on this structure / theme)

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T19:15:13.451Z · LW · GW

Good idea to check Tim Urban's article, Trevor. It seems like he has thought hard on how to make this stuff visual and intuitive and compelling.

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T19:13:16.158Z · LW · GW

Good idea! I could imagine doing something similar with images generated by DALL-E.

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T19:11:36.442Z · LW · GW

[Policy makers]

We don't let companies use toxic chemicals without oversight.

Why let companies use AI without oversight?

[Insert call to action on support / funding for AI governance or regulation]

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T19:07:00.782Z · LW · GW

[Policymakers & ML researchers]

A virus doesn't need to explain itself before it destroys us. Neither does AI.

A meteor doesn't need to warn us before it destroys us. Neither does AI.

An atomic bomb doesn't need to understand us in order to destroy us. Neither does AI.

A supervolcano doesn't need to think like us in order to destroy us. Neither does AI.

(I could imagine a series riffing based on this structure / theme)

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T18:56:26.774Z · LW · GW

[Policy-makers & ML researchers]

In 1901, the Chief Engineer of the US Navy said “If God had intended that man should fly, He would have given him wings.” And on a windy day in 1903, Orville Wright proved him wrong.

Let's not let AI catch-us by surprise.

[Insert call to action]

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T18:53:11.998Z · LW · GW

[Policy makers & ML researchers]

“If a distinguished scientist says that something is possible, he is almost certainly right; but if he says that it is impossible, he is very probably wrong” (Arthur Clarke). In the case of AI, the distinguished scientists are saying not just that something is possible, but that it is probable. Let's listen to them. 

[Insert call to action]

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T18:17:05.420Z · LW · GW

[Policy makers & ML researchers]

If you aren't worried about AI, then either you believe that we will stop making progress in AI or you believe that code will stop having bugs...which is it?

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T18:11:31.860Z · LW · GW

[Policy makers & ML researchers]

“AI doesn’t have to be evil to destroy humanity – if AI has a goal and humanity just happens to come in the way, it will destroy humanity as a matter of course without even thinking about it, no hard feelings” (Elon Musk).

[Insert call to action]

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T18:05:48.557Z · LW · GW

[Tech executives]

If you could not fund that initiative that could turn us all into paperclips...that'd be great.

[Insert call to action]

--

If you could not launch the project that could raise the AI kraken...that'd be great.

[Insert call to action]

--

If you could not build the bot that will treat us the way we treat ants...that'd be great.

[Insert call to action]

--

(I could imagine a series riffing based on this structure / theme)

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T17:59:27.853Z · LW · GW

[ML researchers]

Given that we can't agree on whether a hotdog is a sandwich or not...We should probably start thinking about how to tell a computer what is right and wrong.

[Insert call to action on support / funding for AI governance / regulation etc.]

-

Given that we can't agree on whether a straw has two holes or one...We should probably start thinking about how to explain good and evil to a computer.

[Insert call to action on support / funding for AI governance / regulation etc.]

(I could imagine a series riffing based on this structure / theme)

Comment by jcp29 on [$20K in Prizes] AI Safety Arguments Competition · 2022-04-27T17:56:22.367Z · LW · GW

[ML researchers]

"We're in the process of building some sort of god. Now would be a good time to make sure it's a god we can live with" (Sam Harris, 2016 Ted Talk).