Building Something Smarter

post by Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2008-11-02T17:00:00.000Z · LW · GW · Legacy · 57 comments

Contents

57 comments

Previously in seriesEfficient Cross-Domain Optimization

Once you demystify "intelligence" far enough to think of it as searching possible chains of causality, across learned domains, in order to find actions leading to a future ranked high in a preference ordering...

...then it no longer sounds quite as strange, to think of building something "smarter" than yourself.

There's a popular conception of AI as a tape-recorder-of-thought, which only plays back knowledge given to it by the programmers - I deconstructed this in Artificial Addition, giving the example of the machine that stores the expert knowledge Plus-Of(Seven, Six) = Thirteen instead of having a CPU that does binary arithmetic.

There's multiple sources supporting this misconception:

The stereotype "intelligence as book smarts", where you memorize disconnected "facts" in class and repeat them back.

The idea that "machines do only what they are told to do", which confuses the idea of a system whose abstract laws you designed, with your exerting moment-by-moment detailed control over the system's output.

And various reductionist confusions - a computer is "mere transistors" or "only remixes what's already there" (just as Shakespeare merely regurgitated what his teachers taught him: the alphabet of English letters - all his plays are merely that).

Since the workings of human intelligence are still to some extent unknown, and will seem very mysterious indeed to one who has not studied much cognitive science, it will seem impossible for the one to imagine that a machine could contain the generators of knowledge.

The knowledge-generators and behavior-generators are black boxes, or even invisible background frameworks.  So tasking the imagination to visualize "Artificial Intelligence" only shows specific answers, specific beliefs, specific behaviors, impressed into a "machine" like being stamped into clay.  The frozen outputs of human intelligence, divorced of their generator and not capable of change or improvement.

You can't build Deep Blue by programming a good chess move for every possible position.  First and foremost, you don't know exactly which chess positions the AI will encounter.  You would have to record a specific move for zillions of positions, more than you could consider in a lifetime with your slow neurons.

But worse, even if you could record and play back "good moves", the resulting program would not play chess any better than you do.  That is the peril of recording and playing back surface phenomena, rather than capturing the underlying generator.

If I want to create an AI that plays better chess than I do, I have to program a search for winning moves.  I can't program in specific moves because then the chess player really won't be any better than I am.  And indeed, this holds true on any level where an answer has to meet a sufficiently high standard.  If you want any answer better than you could come up with yourself, you necessarily sacrifice your ability to predict the exact answer in advance - though not necessarily your ability to predict that the answer will be "good" according to a known criterion of goodness.  "We never run a computer program unless we know an important fact about the output and we don't know the output," said Marcello Herreshoff.

Deep Blue played chess barely better than the world's top humans, but a heck of a lot better than its own programmers.  Deep Blue's programmers had to know the rules of chess - since Deep Blue wasn't enough of a general AI to learn the rules by observation - but the programmers didn't play chess anywhere near as well as Kasparov, let alone Deep Blue.

Deep Blue's programmers didn't just capture their own chess-move generator.  If they'd captured their own chess-move generator, they could have avoided the problem of programming an infinite number of chess positions.  But they couldn't have beat Kasparov; they couldn't have built a program that played better chess than any human in the world.

The programmers built a better move generator - one that more powerfully steered the game toward the target of winning game positions.  Deep Blue's programmers surely had some slight ability to find chess moves that aimed at this same target, but their steering ability was much weaker than Deep Blue's.

It is futile to protest that this is "paradoxical", since it actually happened.

Equally "paradoxical", but true, is that Garry Kasparov was not born with a complete library of chess moves programmed into his DNA.  Kasparov invented his own moves; he was not explicitly preprogrammed by evolution to make particular moves - though natural selection did build a brain that could learn.  And Deep Blue's programmers invented Deep Blue's code without evolution explicitly encoding Deep Blue's code into their genes.

Steam shovels lift more weight than humans can heft, skyscrapers are taller than their human builders, humans play better chess than natural selection, and computer programs play better chess than humans.  The creation can exceed the creator.  It's just a fact.

If you can understand steering-the-future, hitting-a-narrow-target as the work performed by intelligence - then, even without knowing exactly how the work gets done, it should become more imaginable that you could build something smarter than yourself.

By building something and then testing it?  So that we can see that a design reaches the target faster or more reliably than our own moves, even if we don't understand how?  But that's not how Deep Blue was actually built.  You may recall the principle that just formulating a good hypothesis to test, usually requires far more evidence than the final test that 'verifies' it - that Einstein, in order to invent General Relativity, must have already had in hand enough evidence to isolate that one hypothesis as worth testing.  Analogously, we can see that nearly all of the optimization power of human engineering must have already been exerted in coming up with good designs to test.  The final selection on the basis of good results is only the icing on the cake.  If you test four designs that seem like good ideas, and one of them works best, then at most 2 bits of optimization pressure can come from testing - the rest of it must be the abstract thought of the engineer.

There are those who will see it as almost a religious principle that no one can possibly know that a design will work, no matter how good the argument, until it is actually tested.  Just like the belief that no one can possibly accept a scientific theory, until it is tested.  But this is ultimately more of an injunction against human stupidity and overlooked flaws and optimism and self-deception and the like - so far as theoretical possibility goes, it is clearly possible to get a pretty damn good idea of which designs will work in advance of testing them.

And to say that humans are necessarily at least as good at chess as Deep Blue, since they built Deep Blue?  Well, it's an important fact that we built Deep Blue, but the claim is still a nitwit sophistry.  You might as well say that proteins are as smart as humans, that natural selection reacts as fast as humans, or that the laws of physics play good chess.

If you carve up the universe along its joints, you will find that there are certain things, like butterflies and humans, that bear the very identifiable design signature and limitations of evolution; and certain other things, like nuclear power plants and computers, that bear the signature and the empirical design level of human intelligence.  To describe the universe well, you will have to distinguish these signatures from each other, and have separate names for "human intelligence", "evolution", "proteins", and "protons", because even if these things are related they are not at all the same.

57 comments

Comments sorted by oldest first, as this post is from before comment nesting was available (around 2009-02-27).

comment by Brian_Jaress2 · 2008-11-02T18:04:09.000Z · LW(p) · GW(p)
To describe the universe well, you will have to distinguish these signatures from each other, and have separate names for "human intelligence", "evolution", "proteins", and "protons", because even if these things are related they are not at all the same.

Speaking of separate names, I think you shouldn't call this "steering the future" stuff "intelligence." It sounds very useful, but almost no one except you is referring to it when they say "intelligence." There's some overlap, and you may feel that what you are referring to is more useful than what they are referring to, but that doesn't give you license to take over the word.

I know you've written a bunch of articles justifying your definition. I read them. I also read the articles moaning that no one understands what you mean when you say "intelligence." I think it's because they use that word to mean something else. So maybe you should just use a different word.

In fact, I think you should look at generalized optimization as a mode of analysis, rather than a (non-fundamental) property. Say, "Let's analyze this in terms of optimization (rather than conserved quantities, economic cost/benefit, etc.)" not, "Let's measure its intelligence."

In one of your earlier posts, people were saying that your definitions are too broad and therefore useless. I agree with them about the broadness, but I think this is still a useful concept if it is taken to be a one of many ways of looking at almost anything.

comment by Caledonian2 · 2008-11-02T19:19:22.000Z · LW(p) · GW(p)
There are those who will see it as almost a religious principle that no one can possibly know that a design will work, no matter how good the argument, until it is actually tested.

It's not necessarily a religious principle, although like anything else, it can be made one. It's a simple truth.

There is a non-trivial distinction between believing that a design will work, and believing that a design is highly likely to work. Maintaining the awareness that our maps do not constrain the territory is hard work, but necessary for a rationalist.

comment by Will_Pearson · 2008-11-02T19:41:11.000Z · LW(p) · GW(p)
Analogously, we can see that nearly all of the optimization power of human engineering must have already been exerted in coming up with good designs to test. The final selection on the basis of good results is only the icing on the cake. If you test four designs that seem like good ideas, and one of them works best, then at most 2 bits of optimization pressure can come from testing - the rest of it must be the abstract thought of the engineer.

Why are we vastly better at optimization (on some problems) than our ancestors if we share the same psychology? Culture.

How is worthwhile scientific culture built up? Testing. Your car designer relies on people testing the properties of materials such as toxicity, tensile strength, melting point, chemical activity and many more. He also relies on the tests of the values of air resistance, wind tunnel simulation accuracy, gravity, likely surface friction and many many more.

The car design is built upon a mountain of tests, not many done by himself. Abstract thought can't get you very far with no tests at all.

comment by Carey_Underwood · 2008-11-02T19:59:18.000Z · LW(p) · GW(p)

Brian, if this definition is more useful, then why isn't that license to take over the term? More specifically, how can there be any argument on the basis of some canonical definition, while the consensus seems that we really don't know the answer yet?

It seems akin to arguing that aerodynamics isn't an appropriate basis for the definition of 'flight', just because a preconceived notion of flight includes the motion of the planets as well as that of the birds, even though the mechanisms turn out to be very different.

comment by Felix2 · 2008-11-02T20:02:03.000Z · LW(p) · GW(p)

It sounds like you're pegging "intelligence" to mean what I'd call a "universal predictor". That is, something that can predict the future (or an unknown) given some information. And that it can do so given a variety of types of sets of unknowns, where "variety" involves more than a little hand-waving.

Therefore, something that catches a fly ball ("knowing" the rules of parabolic movement) can predict the future, but is not particularly "intelligent" if that's all it can do. It may be even a wee bit more "intelligent" if it can also predict where a mortar shell lands. It is even more "intelligent" if it predicts how to land a rocket on the moon. It is even more "intelligent" if it predicts the odds that any given cannon ball will land on a fort's walls. Etc.

I agree with Brain that this is a narrow definition of "intelligence". But that doesn't stop it from being an appropriate goal for AI at this time. That the word, "intelligence" is chosen to denote this goal seems more a result of culture than anything else. AI people go through a filter that extols "intelligence". So ... (One is reminded of many years ago when some AI thinkers had the holy grail of creating a machine that would be able to do the highest order of thinking the AI thinkers could possibly imagine: proving theorems. Coincidently, this is what these thinkers did for a living.)

Here's a thought on pinning down that word, "variety".

First, it seems to me that a "predictor" can be optimized to predict one thing very well. Call it a "tall" predictor (accuracy in Y, problem-domain-ness in X) Or it can be built to predict a lot of things rather poorly, but better than a coin. Call it a "flat" predictor. The question is: How efficient is it? How much prediction-accuracy comes out of this "predictor" given the resources it consumes? Or, using the words, "tall" and "flat" graphically, what's the surface area covered by the predictor, given a fixed amount of resources?

Would not "intelligence", as you mean it, be slightly more accurately defined as how efficient a predictor is and, uh, it's gotta be really wide or we ignore it?

comment by pdf23ds · 2008-11-02T20:12:54.000Z · LW(p) · GW(p)

"Speaking of separate names, I think you shouldn't call this 'steering the future' stuff 'intelligence.'"

I believe he generally calls it "an optimization process". His definition of intelligence was given in an earlier post, and puts more constraints on what counts.

comment by pdf23ds · 2008-11-02T20:14:28.000Z · LW(p) · GW(p)

Brian, Felix seems to understand the distinction Eliezer is making.

comment by Tim_Tyler · 2008-11-02T20:31:10.000Z · LW(p) · GW(p)

I regularly face the problem of wanting to refer to the process of selection acting on random mutations by some simple term. Clearly the term "evolution" won't do - since that term is already taken by a definition which fails to exclude processes such as intelligent design by genetic engineers. Probably the single best solution I have come up with is "primitive evolution".

comment by Brian_Jaress2 · 2008-11-02T21:21:20.000Z · LW(p) · GW(p)
Brian, if this definition is more useful, then why isn't that license to take over the term?

Carey, I didn't say it was a more useful definition. I said that Eliezer may feel that the thing being referred to is more useful. I feel that money is more useful than mud, but I don't call my money "mud."

More specifically, how can there be any argument on the basis of some canonical definition, while the consensus seems that we really don't know the answer yet?

I'm not arguing based on a canonical definition. I agree that we don't have a precise definition of intelligence, but we do have a very rough consensus on particular examples. That consensus rejects rocks, trees, and apple pies as not intelligent. It also seems to be rejecting paperclip maximizers and happy-face tilers.

It seems akin to arguing that aerodynamics isn't an appropriate basis for the definition of 'flight', just because a preconceived notion of flight includes the motion of the planets as well as that of the birds, even though the mechanisms turn out to be very different.

I've never heard anyone say a planet was flying, except maybe poetically. Replace "planets" with "balloons" and it'll get much closer to what I'm thinking.

Replies from: Kenny
comment by Kenny · 2013-06-28T12:56:48.187Z · LW(p) · GW(p)

It does not seem to me that 'intelligence' excludes ("rejects") paperclip maximizers and happy-face tilers.

Does it seem to some that the limited goals of some (hypothetical) beings necessarily prevents them from being intelligent? Is this a failure of imagination, of seriously considering something that is smarter than humans but much more extremely focused in terms of its goals?

comment by Vladimir_Slepnev · 2008-11-02T21:28:59.000Z · LW(p) · GW(p)

Eliezer, I wanna tell you something that will sound offensive, please keep in mind I'm not trying to offend you...

You're making a religion out of your stuff.

Your posts are very different from Robin's - he shows specific applications of rationality, while you preach rationality as a Way. Maybe it has to do with your ethnicity: inventing religions is the #1 unique, stellar specialty of Jews. (Quick examples: Abrahamic religions, socialism, Ayn Rand's view of capitalism. Don't get offended, don't.)

Not saying your personal brand of rationality is wrong - far from it! It's very interesting, and you have taught me much. But as the blog title says, notice and overcome the bias.

Because religions have a way of becoming ugly in the long run.

comment by Caledonian2 · 2008-11-02T21:36:53.000Z · LW(p) · GW(p)

Vladimir, you are slightly incorrect. Eliezer doesn't preach rationality as a Way, he preaches a Way that he claims is rationality.

And like any other preacher, he doesn't take well to people questioning and challenging him. You're likely to see your comments censored.

comment by pdf23ds · 2008-11-02T22:22:19.000Z · LW(p) · GW(p)

I agree that we don't have a precise definition of intelligence, but we do have a very rough consensus on particular examples. That consensus rejects rocks, trees, and apple pies as not intelligent. It also seems to be rejecting paperclip maximizers and happy-face tilers.

I disagree. "Superintelligent" (i.e. really powerful) paperclip maximizers are intelligent.

comment by pdf23ds · 2008-11-02T22:24:15.000Z · LW(p) · GW(p)

Your posts are very different from Robin's - he shows specific applications of rationality, while you preach rationality as a Way.

Eliezer's posts have a specific, mystical, reverent tone that I've seen in many popular science books. Was Carl Sagan hawking a religion?

comment by Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2008-11-02T22:34:48.000Z · LW(p) · GW(p)

Vladimir, you haven't been reading this blog for long, have you? Your comment history shows comments on only recent posts. I've done plenty of application to specifics - see e.g. the Quantum Physics Sequence. What I apply to those specifics is, of course, the Way.

comment by Nick_Tarleton · 2008-11-02T22:38:56.000Z · LW(p) · GW(p)

Pseudo-religious style is sort of a nerd tradition as well.

comment by JamesAndrix · 2008-11-02T23:09:45.000Z · LW(p) · GW(p)

If we want a robot that can navigate mazes, we could put some known pathfinding/search algorithms into it.

Or we could put a neural network in it and run it through thousands of trials with slowly increasing levels of difficulty.

In the first case it is engineered-reliable, in the second case it is learning.

On rereading the opening: "searching possible chains of causality, across learned domains, in order to find actions leading to a future ranked high in a preference ordering"

I feel like you skipped over the concept of learning. Yes we can build maze and chess searchers, because we have learned the domain.

How do we engineer something to be reliably better than us at learning? (To the point that we trust it to use it's learning to construct it's preference ordering.)

comment by PK · 2008-11-03T00:00:12.000Z · LW(p) · GW(p)

"You can't build build Deep Blue by programming a good chess move for every possible position."

Syntax error: Subtract one 'build'.

comment by Caledonian2 · 2008-11-03T00:57:31.000Z · LW(p) · GW(p)

Was Carl Sagan hawking a religion?
Yes. He was trying to convince people that rationality could substitute itself for mysticism, which they highly value.

Pretty much everyone trying to "sell" an idea dip into religious memes, sooner or later -- as my now-deleted comment makes clear.

comment by Felix · 2008-11-03T02:29:31.000Z · LW(p) · GW(p)

@James: If we want a robot that can navigate mazes, we could put some known pathfinding/search algorithms into it. Or we could put a neural network in it and run it through thousands of trials with slowly increasing levels of difficulty.

That evokes some loopy thinking. To wit:

It's always seemed that AI programs, striving for intelligence, can have their intelligence measured by how easy it is to get them to do something. E.g. It's easier to simply run that neural net through a bunch of trials than it is to painstakingly engineer an algorithm for a particular search problem.

So, does that mean that the definition of "intelligence" is: "How easy it is for me to get the intelligent being to do my bidding multiplied by the effect of their actions?"

Or is that a definition of "intelligence we want"? And the definition of "intelligence" is: "The ability to create "intelligence we want" and avoid "intelligence we don't want"?

comment by pdf23ds · 2008-11-03T04:47:37.000Z · LW(p) · GW(p)

And the definition of "intelligence" is: "The ability to create "intelligence we want" and avoid "intelligence we don't want"?

While a recursive definition of intelligence might please Hofstadter, I'd rather avoid it.

comment by JamesAndrix · 2008-11-03T05:00:11.000Z · LW(p) · GW(p)

@Felix I started out trying to make a case that the systems we have that seem most promising are the ones that learn, in one way or another. A general intelligence has to be something that you dump training data into.

It IS easier to run the neural net through trials when you don't understand the problem and might not even know where to start. As eliezer said "Deep Blue's programmers had to know the rules of chess - since Deep Blue wasn't enough of a general AI to learn the rules by observation"

On the one hand, this is fine, we know the rules of physics close enough for the AI to design things we ask for. But it can't do real autonomous planning or self modification (and still be friendly) without a friendly preference ordering/utility function.

The way Eliezer suggests (if I understand) to create a friendly utility function/preference ordering is to have the AI analyze human brains and prefer outcomes we would prefer if we were that smart. But since we don't have the 'rules of the Brain', the AI has to learn. (unless it does a full brain simulation every time it needs to make a preference decision, but I'd rank that as a low preference.)

I'm having trouble imagining a way to engineer a learning system where we know what will come out of it without ourselves understanding the problem. True, we only need to know the end result, and not intermediate steps. But how can we know if a learning system will miss something vital on a problem that is also beyond our scope?

If you didn't know the rules of chess, you could make an AI learn the rules. But when do you bet your life that it hasn't missed something obscure but important?

comment by Vladimir_Slepnev · 2008-11-03T07:28:04.000Z · LW(p) · GW(p)

Vladimir, you haven't been reading this blog for long, have you?

Eliezer, I've lurked here for about a year. The quantum sequence was great (turned me on to many-worlds), but already pretty religious, e.g. the rationale or "it came time to break your allegiance to Science". I ate the tasty intellectual parts and mentally discarded the nasty religious parts. (For example, attacking science by attacking the Copenhagen interpretation was pretty low - most physicists don't even consider interpretations science.) Your recent posts however are all nasty, no tasty. Talmudic.

Thanks for reminding about "Is Humanism A Religion-Substitute?", it's a perfect example of what I'm talking about. You seem to be instinctively religious - you want to worship something, while e.g. for me it's just distasteful.

Religions don't go bad because they are false and stupid. Religions go bad because they live on the "ought" side of is/ought, where there is no true and false. (Cue your morality sequence.)

comment by Matthew_C.2 · 2008-11-03T12:51:25.000Z · LW(p) · GW(p)

Vladimir,

Excellent comments and insight.

The biggest problem with being religious, in the way you have defined, is that your "oughts" start to influence your perception of what "is". We see exactly the same thing happening with traditional religions, with political religions, and here with Singularitarian religion.

The one true worthy action is to perceive the "is" without blurring our vision with the distorting filters of "oughts". This is the true "Way". But, of course, it is not nearly as popular as promising salvation, whether religious or technological.

comment by Lake · 2008-11-03T13:04:16.000Z · LW(p) · GW(p)

For what it's worth, I still find the mock-mysticism stuff fairly entertaining. Hard to think of another running joke that would stay good over the same duration.

Lest we forget: http://lesswrong.com/static/imported/2008/03/27/elimonk2darker.jpg

comment by luzr · 2008-11-03T13:07:17.000Z · LW(p) · GW(p)

Interesting... Reading this entry somewhat awakened my interest in computer chess... And I think I have stumled upon a nice example "inteligence as book smart" misconception.

http://en.wikipedia.org/wiki/Chess960

"Fischer believed that eliminating memorized book moves would level the playing field; as an accidental consequence, it makes computer chess programs much weaker, as they rely heavily on the opening book to beat humans."

Classic example of "bias". The reality:

"In 2005, the chess program Shredder, developed by Stefan Meyer-Kahlen from Düsseldorf, Germany, played two games against Zoltán Almási from Hungary; Shredder won 2-0."

That was only the second public match of computer vs human.

It is also worth noting that almost the equal relevance in the wiki article is given to human competitions as is to computer ones.

Now, of course, the remaining question is: "does the machine understand chinesse if it can speak better than chinaman"? :)

comment by Ben_Jones · 2008-11-03T13:15:46.000Z · LW(p) · GW(p)

"If your dreams and aspirations interfere with your rationality and impartiality then you're no longer following the Way. The Way doesn't drive one off course; one rather loses sight of the Way and hence goes off course."

[The Book of Eliezer 4:24 18-20]

comment by Silas · 2008-11-03T14:09:28.000Z · LW(p) · GW(p)

Quick question: How would you build something smarter, in a general sense, than yourself? I'm not doubting that it's possible, I'm just interested in knowing the specific process one would use.

Keep it brief, please. ;-)

comment by luzr · 2008-11-03T14:29:31.000Z · LW(p) · GW(p)

@Silas

Quick reply: I think we will not know exactly, but that does not mean we cannot do it.

More detailed: I believe that the process will go with some sort of connectionist approach maybe mixed with some genetic algorithm voodoo. The key is reaching high complexity system that does something and starts to self-evolve (which in fact is nearly the same thing as learn).

Anyway, I believe that when all is done, we will have the same problem understanding Strong-AI mental processes as we have human brain ones.

comment by Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2008-11-03T19:13:21.000Z · LW(p) · GW(p)
Religions don't go bad because they are false and stupid. Religions go bad because they live on the "ought" side of is/ought, where there is no true and false. (Cue your morality sequence.)

So you're telling me I ought to stop doing that?

comment by Chris_Hibbert · 2008-11-03T19:35:12.000Z · LW(p) · GW(p)

@Silas

I thought the heart of EY's post was here:

even if you could record and play back "good moves", the resulting program would not play chess any better than you do.

If I want to create an AI that plays better chess than I do, I have to program a search for winning moves. I can't program in specific moves because then the chess player really won't be any better than I am. [...] If you want [...] better [...], you necessarily sacrifice your ability to predict the exact answer in advance - though not necessarily your ability to predict that the answer will be "good" according to a known criterion of goodness. "We never run a computer program unless we know an important fact about the output and we don't know the output," said Marcello Herreshoff.

So the heart of the AI is something that can generate and recognize good answers. In game playing programs, it didn't take long for the earliest researchers to come up with move and position evaluators that they have been improving on ever since. There have even been some attempts at general move and position evaluators. (See work on Planner, Micro-Planner, and Conniver, which will probably lead you to other similar work.) Move generation has always been simpler in the game worlds than it would be for any general intelligence. The role of creativity hasn't been explored that much AFAICT, but it's crucial in realms where the number of options at any point are so much larger than in game worlds.

The next breakthrough will require some different representation of reality and of goals, but Eli seems to be pointing at generation and evaluation of action choices as the heart of intelligent behavior. The heart of it seems to be choosing a representation that makes generation and analysis of possible actions tractable. I'm waiting to see if EY has any new ideas on that front. I can't see how progress will be made without it, even in the face of all of EY's other contributions to understanding what the problem is and what it would mean to have a solution.

And EY has clearly said that he's more interested in behavior ("steering the future") than recognition or analysis as a characteristic of intelligence.

comment by Vladimir_Slepnev · 2008-11-03T19:52:10.000Z · LW(p) · GW(p)

So you're telling me I ought to stop doing that?

Cute counter but fallacious IMO. There are systems of oughts that don't look and sound like religions. For example, I don't write sermons for mine. Anyway, you're not engaging my central point, just nitpicking on an illustratory phrase.

Replies from: Kenny
comment by Kenny · 2013-06-28T14:12:49.279Z · LW(p) · GW(p)

What does a "system of oughts" look like? And why would it matter what they sound like? Does the content of the system change if statements about it are read aloud by voices of different pitch?

comment by Tim_Tyler · 2008-11-03T21:12:56.000Z · LW(p) · GW(p)
"We never run a computer program unless we know an important fact about the output and we don't know the output," said Marcello Herreshoff.

That's a fallacy - as I have previously explained on Overcoming Bias: we do not just run computer programs to learn things we also run them to do things.

For example, we use computer programs to automate common tasks. Say I want the numbers from 1 to 100 printed down the left side of a piece of paper. It would be reasonable to write a Python script to do that. Not because there is something unknown about the output, but because a computer and a printer can do it faster, better and neater than I can.

comment by Bob_Mottram · 2008-11-03T23:26:09.000Z · LW(p) · GW(p)

This seems to happen a lot in computer vision - particularly object recognition - where the prevailing paradigm involves training the system on images taken from numerous perspectives which then become exemplars for subsequent matching. The problem with this is that it just doesn't scale. Having a large number of 2D training examples, then trying to match also in 2D, works for objects which are inherently flat - like the cover of a book - but not for objects with significant 3D structure, such as a chair. For more effective recognition the system needs to capture the 3D essence of the object, not it's innumerable 2D shadows.

comment by Kaj_Sotala · 2008-11-04T08:08:59.000Z · LW(p) · GW(p)

Tim Tyler: That's a fallacy - as I have previously explained on Overcoming Bias: we do not just run computer programs to learn things we also run them to do things. ... Not because there is something unknown about the output, but because a computer and a printer can do it faster, better and neater than I can.

That's a silly distinction to use as a basis of criticism. Would you also say that it's a fallacy to claim that we use cars to get to distant places, since we could also get there by walking?

comment by Tim_Tyler · 2008-11-04T18:44:50.000Z · LW(p) · GW(p)
That's a silly distinction to use as a basis of criticism. Would you also say that it's a fallacy to claim that we use cars to get to distant places, since we could also get there by walking?

The claim was that something never happens. Clearly there are counter-examples where it does indeed happen - and indeed a whole class of violations.

The car analogy is not relevant. Better to imagine that it had been claimed that humans never travel anywhere, unless they are in a car - in which case observing people travelling by bicycle would be the corresponding counter-example.

comment by Nick_Tarleton · 2008-11-04T19:49:06.000Z · LW(p) · GW(p)

It's a stretch, but one could say that in that case you want an output with the property [is a physical system consisting of the numbers 1 to 100 prettily printed], and don't know any exact arrangement of atoms with this property.

comment by Silas · 2008-11-04T20:45:48.000Z · LW(p) · GW(p)

Nick_Tarleton: Actually, Tim_Tyler's claim would still be true there, because you may want to print out that list, even if you knew some exact arrangement of atoms with that property.

However, I think Marcello's Rule is still valid there and survives Tim_Tyler's objection: in that case, what you don't know is "the sensation arising from looking at a the numbers 1 through 100 prettily printed". Even if you had seen such a list before, you probably would want to print it out unless your memory were perfect.

My claim generalizes nicely. For example, even if you ran a program for the purpose of automating a farm, and knew exactly how the farm would work, then what you don't know in that case is "the sensation of subsisting for x more days". Although Marcello's Rule starts to sound vacuous at that point.

Hey, make a squirrely objection, get a counterobjection twice as squirrely ;-)

comment by Eliezer Yudkowsky (Eliezer_Yudkowsky) · 2008-11-04T22:11:35.000Z · LW(p) · GW(p)

The truly squirrely response would be to claim that a farm automation program that simply repeats an exact, preprogrammed sequence over and over is a "tape recorder" rather than a "computer program".

comment by Silas · 2008-11-04T22:47:21.000Z · LW(p) · GW(p)

@Eliezer_Yudkowsky: It wouldn't be an exact sequence repeating, since the program would have to handle contingencies, like cows being uncooperative because of insufficiently stimulating conversation.

comment by Tim_Tyler · 2008-11-04T23:09:06.000Z · LW(p) · GW(p)

There seem to be even more objections :-(

What is unknown is not the output, but the sensation it generates;

Checking with the original wording: "We never run a computer program unless we know an important fact about the output and we don't know the output."

It seems to specify that the output is what is unknown - not the sensations that output generates in any particular observer.

An automation program that does something is not a computer program at all;

That seems fairly trivially wrong to me. Plenty of scripts that people write to do things surely are computer programs - by most people's definitions. E.g. they are written in common computer programming languages, and execute on conventional computer hardware.

To "know" the output of a program, it must be physically at hand.

That seems like a rather biblical sense of the word "know" ;-)

Do we "know" the output of a program that prints the square root of 81 before we run it? I think most would say "yes" - even if they did not have the printed output in front of them.

Why so much spirited defense of Marcello Herreshoff's inaccurate statement? ;-)

comment by Cyan2 · 2008-11-04T23:37:00.000Z · LW(p) · GW(p)

For what it's worth, Tim Tyler, I'm with you. Utility scripts count as programs in my books.

comment by Silas · 2008-11-04T23:55:53.000Z · LW(p) · GW(p)

Very worthwhile points, Tim_Tyler.

First of all, the reason for my spirited defense of MH's statement is that looked like a good theory because of how concise it was, and how consistent with my knowledge of programs it was. So, I upped my prior on it and tended to see apparent failures of it as a sign I'm not applying it correctly, and that further analysis could yield a useful insight.

And I think I that belief is turning out to be true:

It seems to specify that the output is what is unknown - not the sensations that output generates in any particular observer.

But the sensations are a property of the output. In a trivial sense: it is a fact about the output, that a human will perceive it in a certain way.

And in a deeper sense, the numeral "9" means "that that someone will perceive as symbol representing the number nine in the standard number system". I'm reminded of Douglas Hofstadter's claim that definition of individual letters is an AI-complete problem because you must know a wealth of information about the cognitive system to be able to identify the full set of symbols someone will recognize as e.g. an "A".

This yields the counterintuitive result that, for certain programs, you must reference the human cognitive system (or some concept isomorphic thereto) in listing all the facts about the output. That result must hold for any program whose output will eventually establish mutual information with your brain.

Am I way off the deep end here? :-/

comment by Tim_Tyler · 2008-11-05T09:00:09.000Z · LW(p) · GW(p)

Further analysis might conceivably yield a useful insight - I'm only suggesting that the claim as it stands is an over-generalisation and has exceptions.

As well as automation programs, we must consider programs that are run to occupy time or burn resources, programs that are run in order to DOS-attack the computer (see The Forbin Project), pointless programs, and so on. In such cases the output may be irrelevant or non-existent, rather than unknown.

comment by Silas · 2008-11-05T17:36:23.000Z · LW(p) · GW(p)

Further analysis, you say, Tim_Tyler? Could you please redirect effort away from putdowns and into finding what was wrong with the reasoning in my previous comment?

comment by Will_Pearson · 2008-11-05T18:17:11.000Z · LW(p) · GW(p)
But the sensations are a property of the output. In a trivial sense: it is a fact about the output, that a human will perceive it in a certain way.

Can I ask you to taboo property.

I like to think there is some difference between a metals melting point and that it is Freds least favourite metal. I consider the second to be a property of Fred (that is how I would code it anyway), not of the metal.

comment by Silas · 2008-11-05T19:12:23.000Z · LW(p) · GW(p)

Okay, fair challenge.

I agree about your metal example, but it differs significantly from my discussion of the list-output program for the non-trivial reason I gave: specifically, the output is defined by its impact on people's cognitive structure.

Look at it this way: Tim_Tyler claims that I know everything there is to know about the output of a program that spits out the integers from 1 to 100. But, when I get the output, what makes me agree that I am in fact looking at those integers? Let's say that when printing it out (my argument can be converted to one about monitor output), I see blank pages. Well, then I know something messed up: the printer ran out of ink, was disabled, etc.

Now, here's where it gets tricky: what if instead it only sorta messes up: the ink is low and so it's applied unevenly so that only parts of the numbers are missing? Well, depending on how badly it messes up, I may or may not still recognize the numbers as being the integers 1-100. It depends on whether it retains enough of the critical characteristics of those numbers for me to so recognize them.

To tie it back to my original point, what this all means is that the output is only defined with respect to a certain cognitive system: that determines whether the numbers are in fact recognizable as 9's, etc. If it's not yet clear what the difference is between this and metal's melting point, keep in mind that we can write a program to find a metal's melting point, but we can't write a program that will look at a printout and know if it retains enough of its form that a human recognizes it as any specific letter -- not yet, anyway.

comment by Will_Pearson · 2008-11-05T19:28:35.000Z · LW(p) · GW(p)
the output is defined by its impact on people's cognitive structure.

Is it literally? Are you saying I couldn't send a message to someone that enabled them to print out a list of the first hundred integers without referencing a human's cognitive structure.

Sure if it went wrong they wouldn't be able to fix it. But even in that case it might be as well to send three different programs of things that do the same thing, so they could fix the incorrect one, rather than sending a full spec.

comment by Tim_Tyler · 2008-11-05T19:39:11.000Z · LW(p) · GW(p)
Could you please redirect effort away from putdowns and into finding what was wrong with the reasoning in my previous comment?

Not a "putdown". The whole idea looks like it needs major surgery to me - at least I can't see much of interest in it as it stands. Think you can reformulate it so it makes sense? Be my guest.

Tim_Tyler claims that I know everything there is to know about the output of a program that spits out the integers from 1 to 100. [snip printer running out of ink]

"Know" should not be read as "know with certainty". We don't know anything with certainty. Anyway, such a criticism cuts against the original claim as well - since that contained "know" as well as "don't know".

comment by Silas · 2008-11-05T21:42:33.000Z · LW(p) · GW(p)

Will_Pearson: Is it literally? Are you saying I couldn't send a message to someone that enabled them to print out a list of the first hundred integers without referencing a human's cognitive structure.

Yes, that's what I'm saying. It's counterintuitive because you so effortlessly refernce others' cognitive structures. In communicating, you assume a certain amount of common understanding, which allows you to know whehter your message will be understood. In sending such a message, you rely on that information. You would have to think, "will they understand what this sentence means", "can they read this font", etc.

Tim_Tyler: The whole idea looks like it needs major surgery to me - at least I can't see much of interest in it as it stands. Think you can reformulate it so it makes sense? Be my guest.

Certainly. All you have to do is read it so you can tell me what about it doesn't make sense.

Anyway, such a criticism cuts against the original claim as well - since that contained "know" as well as "don't know".

Which contests the point how?

comment by Will_Pearson · 2008-11-05T22:16:00.000Z · LW(p) · GW(p)
Yes, that's what I'm saying. It's counterintuitive because you so effortlessly refernce others' cognitive structures. In communicating, you assume a certain amount of common understanding, which allows you to know whehter your message will be understood. In sending such a message, you rely on that information. You would have to think, "will they understand what this sentence means", "can they read this font", etc.

Personally I'm thinking about communicating with machines myself, so I'm thinking something a long the lines of the following.

#include
void main()
{ int x; for( x= 1;x<101; x++){printf("%d\n",x);} }


That completely defines what I want the system to do. I do not reference other peoples thoughts.

Now it does rely on shared knowledge, the system has to have the standard libraries and a c compiler. But I could send that and still not reference humanity at all. I could send a C compiler in machine code. But that relies on me sharing knowledge with the other side on what machine code it uses.

With no shared knowledge, they cannot understand any type of definition I send. No matter if I reference human senses or not.

comment by Genius2008 · 2008-11-06T06:06:00.000Z · LW(p) · GW(p)

This is of course a straw-man claim, computers are merely transistors arranged in patterns that execute algorithms that are pre-defined. The brain also performs basic operations at a neuronal level as well it has been shown that neurons can do a wide range of operations on a given input. The real question is what is the algorithm that is being executed and how does the brain know which algorithm to use or which operation to execute? And of course the brain can change its wiring patterns making things even more interesting.

"Deep Blue played chess barely better than the world's top humans, but a heck of a lot better than its own programmers. Deep Blue's programmers had to know the rules of chess - since Deep Blue wasn't enough of a general AI to learn the rules by observation - but the programmers didn't play chess anywhere near as well as Kasparov, let alone Deep Blue."

A feeble example since chess is a mathematical exercise that readily lends itself to mathematical analysis and algorithm development. Human intelligence while more predictable then some like to believe, still is not nearly as predictable as chess. Lets be real honest about Deep Blue the real advantage over a person is that Deep Blue can analyze far more moves then a person can and can plan further ahead. If we change the game to something like facial recognition the computer is far worse for the most part though facial recognition has been improving.

"If you can understand steering-the-future, hitting-a-narrow-target as the work performed by intelligence - then, even without knowing exactly how the work gets done, it should become more imaginable that you could build something smarter than yourself."

Key word imaginable, not necessarily possible since the problem of AGI is not nearly as simple as Chess. Nor does a good chess computer make even the slightest case for AGI. AGI is only proven through construction the rest is babble and hollow claims. It is akin to claiming to be able to build a house all by yourself and then never really doing it, to make that claim one must prove it by the doing.

"There are those who will see it as almost a religious principle that no one can possibly know that a design will work, no matter how good the argument, until it is actually tested."

What works in theory and what works in praxis are two different things sometimes, since your not an engineer you wouldn't know. Try building a design and you will understand until then you really are an outsider.

"Just like the belief that no one can possibly accept a scientific theory, until it is tested. But this is ultimately more of an injunction against human stupidity and overlooked flaws and optimism and self-deception and the like - so far as theoretical possibility goes, it is clearly possible to get a pretty damn good idea of which designs will work in advance of testing them."

Never the less proof through the testing of the idea is needed no matter how good the theory, without testing you have nothing. The point of testing is to ensure that the theory works in practice the same way as it does on paper.

Sorry for cutting out swaths of your post, they were swaths of verbiage not worth the answering. Please if your going to speak about these thing try engineering a real AI and then come back to these philosophical meanderings. Until you have real experience in the programming and the math of AI, beyond Bayesian Statistics you cannot hope to make a case for AGI or build one yourself.

comment by Tim_Tyler · 2008-11-07T18:51:00.000Z · LW(p) · GW(p)

Genius2008's post seems a bit antagonistic. Also, it contains the idea that "AGI is only proven through construction the rest is babble and hollow claims".

We already have intelligent agents on the planet - surely an excellent and highly practical demonstration of the viability of the enterprise.

comment by Genius2008 · 2008-11-08T19:49:00.000Z · LW(p) · GW(p)

I am sorry Tim, that I am not properly respectful of your hero Eliezer Yudkowsky or the AGI fields claims. The fact that humans are intelligent in no way proves AGI. If you can't comprehend that AGI is an engineering problem then you really are out of touch with what is going on. The only way to prove that AGI is possible is to build it until then its just a belief nothing more there is no necessity attached to it.

Replies from: timtyler
comment by timtyler · 2009-09-20T16:10:12.870Z · LW(p) · GW(p)

We will build very intelligent machines - I take that for granted. If you think otherwise, fine - but don't expect a debate on the issue from me. Nor does this seem like the right place for such a discussion. Perhaps try comp.ai.philosophy - they seem to like such banter over there.