Posts

Comments

Comment by rlsj on Superintelligence 24: Morality models and "do what I mean" · 2015-02-26T22:00:33.643Z · LW · GW

"[Y]ou notice that [a proposed action] has some harmful effect in the future on someone you consider morally important. You then end up not being able to do anything ..."

Not being able to do that thing, yes, and you shouldn't do it -- unless you can obviate the harm. A case in point is the AGI taking over management of all commodity production and thus putting the current producers out of work. But how is that harmful to them? They can still perform the acts if they wish. They can't earn a living, you say? Well, then, let the AGI support them. Ah, but then, you suppose, they can't enjoy the personal worth that meaningful employment reinforces. The what? Let's stick to the material, please.

"You then end up not being able to do nothing -- because doing nothing also leads to harm in the future."

That does not follow. Doing nothing is always an option under the brass rule. Morally you are not the cause of any harm that then occurred, if any.

Commission vs. omission [of causative actions]: Omitting an action may indeed allow an entity to come to harm, but this is not a moral issue unless acting would harm that entity or another, perhaps to a lesser degree. Commission -- taking action -- is the problematic case. I repeat: a coded moral system should be restrictive, not promotional. Preventing external harm may be desirable and admirable but is never morally imperative, however physically imperative it may be.

"[Not adding a safety feature] does make the system-designer responsible for [the resulting] accidents in some way ..."

Only by the "Good Samaritan" moral code, in which this society is so dolefully steeped. I prefer Caveat emptor. It may be that when AGIs are the principal operators of harmful equipment, the obsession with safety will moderate.

Comment by rlsj on Superintelligence 24: Morality models and "do what I mean" · 2015-02-25T02:34:59.914Z · LW · GW

CEV, MR, MP ... We do love complexity! Is such love a defining characteristic of intelligent entities?

The main point of all morality, as it is commonly practiced and understood, is restrictive, not promotional. A background moral code should not be expected to suggest goals to the AI, merely to denigrate some of them. The Libertarian "brass" rule is a case in point: "Do not unto others as you would not have them do unto you," which may be summarized as "Do no harm."

Of course, "others" has to be defined, perhaps as entities demonstrating sufficiently complex behavior, and exceptions have to be addressed, such as a third-party about to harm a second party. Must you restrain the third-party and likely harm her instead?

"Harm" will also need precise definition but that should be easier.

The brass rule does not require rendering assistance. Would ignoring external delivery of harm be immoral? Yes, by the "Good Samaritan" rule, but not by the brass rule. A near-absolute adherence to the brass rule would solve most moral issues, whether for AI or human.

"Near-absolute" because all the known consequences of an action must be considered in order to determine if any harm is involved and if so, how negatively the harm weighs on the goodness scale. An example of this might be a proposal to dam a river and thereby destroy a species of mussel. Presumably mussels would not exhibit sufficiently complex behavior in their own right, so the question for this consequence becomes how much their loss would harm those who do.

Should an AI protect its own existence? Not if doing so would harm a human or another AI. This addresses Asimov's three laws, even the first. The brass rule does not require obeying anything.

Apart from avoiding significant harm, the selection of goals does not depend on morality.

--rLsj

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-17T23:42:19.523Z · LW · GW

How does "just following orders," a la Nuremberg, bear upon this issue? It's out of control when its behavior is neither ordered nor wanted.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T20:39:49.594Z · LW · GW

How would you tell? By its behavior: doing something you neither ordered nor wanted.

Think of the present-day "autonomous laborer" with an IQ about 90. The only likely way to lose control of him is for some agitator to instill contrary ideas. Censorship for robots is not so horrible a regime.

Who is it that really wants AGI, absent proof that we need it to automate commodity production?

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T19:56:58.553Z · LW · GW

Excuse me? What makes you think it's in control? Central Planning lost a lot of ground in the Eighties.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T19:49:38.006Z · LW · GW

Please, Madam Editor: "Without the benefit of hindsight," what technologies could you possibly expect?

The question should perhaps be, What technology development made the greatest productive difference? Agriculture? IT? Et alia? "Agriculture" if your top appreciation is for quantity of people, which admittedly subsumes a lot; IT if it's for positive feedback in ideas. Electrification? That's the one I'd most hate to lose.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T19:30:57.135Z · LW · GW

After achieving a society of real abundance, further economic growth will have lost incentive.

We can argue whether or not such a society is truly reachable, even if only in the material sense. If not, because of human intractability or AGI inscrutability, progress may continue onward and upward. Perhaps here, as in happiness, it's the pursuit that counts.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T03:05:08.424Z · LW · GW

An AI can be dangerous only if it escapes our control. The real question is, must we flirt with releasing control in order to obtain a necessary or desirable usefulness? It seems likely that autonomous laborers, assembly-line workers, clerks and low-level managers would, without requiring such flirtation, be useful and sufficient for the society of abundance that is our main objective. But can they operate without a working AGI? We may find out if we let the robots stumble onward and upward.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T02:45:10.911Z · LW · GW

"Does this mean that society would never discover certain facts had the most brilliant people not existed?"

Absolutely! If they or their equivalent had never existed in circumstances of the same suggestiveness. My favorite example of this uniqueness is the awesome imagination required first to "see" how stars appear when located behind a black hole -- the way they seem to congregate around the event horizon. Put another way: the imaginative power able to propose star deflections that needed a solar eclipse to prove.

Comment by rlsj on Superintelligence Reading Group - Section 1: Past Developments and Present Capabilities · 2014-09-16T01:40:36.702Z · LW · GW

For anything whose function and sequencing we thoroughly understand the programming is straightforward and easy, at least in the conceptual sense. That covers most games, including video games. The computer's "side" in a video game, for example, which looks conceptually difficult, most of the time turns out logically to be only decision trees.

The challenge is the tasks we can't precisely define, like general intelligence. The rewarding approach here is to break down processes into identifiable subtasks. A case in point is understanding natural languages, one of whose essential questions is, "What is the meaning of "meaning?" In terms of a machine it can only be the content of a subroutine or pointers to subroutines. The input problem, converting sentences into sets of executable concepts, is thus approachable. The output problem, however, converting unpredictable concepts into words, is much tougher. It may involve growing decision trees on the fly.