Open Problems Related to the Singularity (draft 1)

post by lukeprog · 2011-12-13T10:57:33.916Z · LW · GW · Legacy · 42 comments

Contents

  Safe AI Architectures
  Safe AI Goals
  Strategy
None
42 comments

"I've come to agree that navigating the Singularity wisely is the most important thing humanity can do. I'm a researcher and I want to help. What do I work on?"

The Singularity Institute gets this question regularly, and we haven't published a clear answer to it anywhere. This is because it's an extremely difficult and complicated question. A large expenditure of limited resources is required to make a serious attempt at answering it. Nevertheless, it's an important question, so we'd like to work toward an answer.

A few preliminaries:

Next, a division of labor into "problem categories." There are many ways to categorize the open problems; some of them are probably more useful than the one I've chosen below.

The list of open problems below is very preliminary. I'm sure there are many problems I've forgotten, and many problems I'm unaware of. Probably all of the problems are stated relatively poorly: this is only a "first step" document. Certainly, all listed problems are described at an extremely "high" level, very far away (so far) from mathematical precision, and can be broken down into several and often dozens of subproblems.

Safe AI Architectures

Safe AI Goals

Strategy


My thanks for some notes written by Eliezer Yudkowsky, Carl Shulman, and Nick Bostrom, from which I've drawn.

42 comments

Comments sorted by top scores.

comment by XiXiDu · 2011-12-13T12:33:36.517Z · LW(p) · GW(p)

From my, arguably layman, perspective it seems that making progress on a lot of those problems makes unfriendly AI more probable as well. If for example you got an ideal approximation of perfect Bayesianism, this seems like something that could be used to build any sort of AGI.

Replies from: lukeprog
comment by lukeprog · 2011-12-13T12:41:48.528Z · LW(p) · GW(p)

Not literally "any sort of AGI" of course, but... yes, several of the architecture problems required for FAI also make uFAI more probable. Kind of a shitty situation, really.

Replies from: Technoguyrob
comment by robertzk (Technoguyrob) · 2011-12-18T09:56:23.414Z · LW(p) · GW(p)

Wikipedia says Steve Ohomundro has "discovered that rational systems exhibit problematic natural 'drives' that will need to be countered in order to build intelligent systems safely."

Is he referring to the same problem?

EDIT: I answered my question by finding this.

comment by [deleted] · 2011-12-13T19:11:21.309Z · LW(p) · GW(p)

This is the best thing ever, Luke. Thanks for your work.

comment by [deleted] · 2011-12-15T01:08:14.837Z · LW(p) · GW(p)

What is the ideal theory of how to handle logical uncertainty?

Some possibly relevant work:

Replies from: None
comment by [deleted] · 2011-12-15T06:55:57.988Z · LW(p) · GW(p)

.

Replies from: lukeprog
comment by lukeprog · 2011-12-16T01:02:54.757Z · LW(p) · GW(p)

Thanks!

Replies from: None
comment by [deleted] · 2011-12-16T04:48:50.697Z · LW(p) · GW(p)

My next task was going to be summarizing logical uncertainty insights made in lesswrong comments and posts, but I found Wei Dai's list of resources, which led to a new search of academic literature. My reading list, in decreasing order of importance, now looks like:

  • Bayesian Networks for Logical Reasoning by Williamson
  • Unifying Logical and Probabilistic Reasoning by Haenni
  • Recursive Causality in Bayesian Networks and Self-Fibring Networks by Williamson and Gabbay
  • Possible Semantics for a Common Framework of Probabilistic Logics by Haenni, Romeijn, Wheeler, and Williamson
  • Non-deductive Logic in Mathematics by Franklin
  • A Derivation of Quasi-Bayesian Theory by Cozman
  • Decision Theory without Logical Omniscience by Lipman
  • Slightly More Realistic Personal Probability by Hacking
comment by thomblake · 2011-12-14T00:24:41.999Z · LW(p) · GW(p)

An important step towards answering "What are the most important problems in your field, and why aren't you working on them?"

comment by Shmi (shminux) · 2011-12-14T03:46:20.032Z · LW(p) · GW(p)

Our research priorities are unclear.

Editing and publishing any existing novel results, provided EY and others have some, should the be the top priority. Surely there is an arxiv for that.

Replies from: JoshuaFox
comment by JoshuaFox · 2011-12-14T09:48:44.262Z · LW(p) · GW(p)

"existing novel results, provided EY and others have some"

Indeed there are. TDT, for example, has not yet received an academic writeup. There are lots of ideas scattered through LW which could be published in journals. And the great thing about academic writing is that you are allowed to use other people's ideas, as long as you cite them. You are considered to be doing them a favor when you do that.

In general, this means that one sprinkles another person's ideas within one's own analysis; if a direct rewrite of, e.g., the TDT paper, for a journal is intended, then the original non-academic author should get credit as co-author.

I understand the point that it might not be worth the time of EY or other SI Fellows to publish ideas in journals. But if some lesser lights want to contribute, they can so so in this way.

Replies from: shminux, None
comment by Shmi (shminux) · 2011-12-14T17:55:59.454Z · LW(p) · GW(p)

I understand the point that it might not be worth the time of EY or other SI Fellows to publish ideas in journals.

One can always post a paper on the arxiv.org preprint server, without going through a peer-review process first. Presumably, one of the CoRR subsections would be appropriate. This is always worth the time spent.

comment by [deleted] · 2011-12-14T13:23:18.163Z · LW(p) · GW(p)

It would a be a breach of research ethics for some "lesser light" (really?) to merely rewrite the TDT paper, add Yudkowsky as a coauthor, and publish it. At minimum, to qualify for coauthorship, Yudkowsky would have to review and approve the draft, and that process could take an indefinite amount of time.

Anything else would still be at worst plagiarism, and at best fradulent authorship.

Replies from: JoshuaFox
comment by JoshuaFox · 2011-12-14T14:27:26.569Z · LW(p) · GW(p)

Certainly, EY would have to serve as a coauthor if the published article was closely based on the original, and of course he would have to agree to that.

But I think that coauthorship is a less likely scenario, and the first idea I mentioned--- use of certain key ideas with citation -- is a more likely one.

Replies from: shminux
comment by Shmi (shminux) · 2011-12-14T17:53:39.050Z · LW(p) · GW(p)

Citing an existing write-up in your own research paper adding something new to the TDT, or in an explicitly educational/expository non-research paper is OK. Rewriting existing ideas as your own research paper is not.

Replies from: JoshuaFox
comment by JoshuaFox · 2011-12-14T20:23:24.651Z · LW(p) · GW(p)

Of course. My original comment was meant to convey, through the words "citations," and "coauthor," that proper credit must always be given.

comment by [deleted] · 2011-12-13T19:06:28.428Z · LW(p) · GW(p)

"Lob" needs an umlaut, or else to be written Loeb.

"How can this be process be approximated computably and tractably?"

comment by lukeprog · 2011-12-14T01:31:51.738Z · LW(p) · GW(p)

Does anyone have a superior problem categorization? Does anyone have a correction or addition to make?

Replies from: ESRogs, timtyler
comment by ESRogs · 2011-12-14T02:48:09.153Z · LW(p) · GW(p)

Minor editorial comments:

Consider expanding WBE the first time it is mentioned. I'm a regular reader and couldn't think of what it referred to until I searched the site.

I believe "ok" should be either "OK" or "okay".

comment by timtyler · 2011-12-14T11:22:21.533Z · LW(p) · GW(p)

The list had me wondering where the political problems went.

Replies from: XiXiDu
comment by XiXiDu · 2011-12-14T13:52:33.426Z · LW(p) · GW(p)

The list had me wondering where the political problems went.

You're right. If at some point the general public starts to take risks from AI seriously and realizes that SI is actually trying to take over the universe without their consensus then a better case scenario will be that SI gets closed and its members send to prison. Some of the not so good scenarios might include the complete extermination of the Bay Area if some foreign party believes that they are close to launching an AGI capable of recursive self-improvement.

Sounds ridiculous? Well, what do you think will be the reaction of governments and billions of irrational people who learn and actually believe that a small group of American white male (Jewish) atheist geeks is going to take over the whole universe? BOOM instead of FOOM.

Reference:

...—though it may be an idealistic dream—I intend to plunge into the decision theory of self-modifying decision systems and never look back. (And finish the decision theory and implement it and run the AI, at which point, if all goes well, we Win.)

Eliezer Yudkowsky in an interview with John Baez.

Replies from: timtyler
comment by timtyler · 2011-12-14T14:55:04.844Z · LW(p) · GW(p)

If at some point the general public starts to take risks from AI seriously and realizes that SI is actually trying to take over the universe without their consensus then a better case scenario will be that SI gets closed and its members send to prison.

It doesn't sound terribly likely. People are more likey to guffaw: So: you're planning to take over the world? And you can't tell us how because that's secret information? Right. Feel free to send us a postcard letting us know how you are getting on with that.

Well, what do you think will be the reaction of governments and billions of irrational people who learn and actually believe that a small group of American white male (Jewish) atheist geeks is going to take over the whole universe?

Again, why would anyone believe that, though? Plenty of people dream of ruling the universe - but - so far, nobody has pulled it off.

Most people are more worried about the secret banking cabal with the huge supercomputers, the billions of dollars in spare change and the shadowy past - who are busy banging on the core problem of inductive inference - than they are about the 'friendly' non-profit with its videos and PDF files - and probably rightfully so.

comment by JoshuaFox · 2011-12-13T11:29:13.138Z · LW(p) · GW(p)

This is great!

One path to advancing research is to take advantage of some low-hanging fruit for mainstream research: A variety of problems in existing academic areas. It might be relatively easier to get people who are working "in the system" to get started on these. For example, reflective decision theory.

Replies from: lukeprog, JoshuaFox
comment by lukeprog · 2011-12-13T11:33:49.606Z · LW(p) · GW(p)

Yes exactly. Same thing with value extrapolation algorithms (aka 'ideal preference' or 'full information' theories of value; see Muehlhauser & Helm 2011.)

comment by JoshuaFox · 2011-12-13T13:51:40.333Z · LW(p) · GW(p)

Another example: You could discuss many questions in psychology or the philosophy of mind asking how the specifically human aspects differ from what could be found in minds-in-general. This is well-defined enough to be discussed intelligently in a term paper.

(Such discussions in behavioral economics often compare humans to perfect rational agents; in ev.psych, the adaptive value of human psychological features are described. But rarely is the universe of minds under consideration explicitly expanded beyond the human.)

comment by lukeprog · 2011-12-14T02:18:31.155Z · LW(p) · GW(p)

Why are my spaces suddenly collapsed in the OP? I checked; there aren't any

elements.

EDIT: 'twas a bug that was fixed, by Trike.

comment by FiftyTwo · 2011-12-14T02:13:03.173Z · LW(p) · GW(p)

Given that finding solutions to all other problems becomes exponentially easier the more people are involved, then surely our highest priority should be awareness spreading, recruitment and the 'Strategy' goals related to that.

Replies from: FAWS
comment by FAWS · 2011-12-14T13:01:42.952Z · LW(p) · GW(p)

But being seen as having done impressive work already is one way to make your word heard and probably one of the more powerful recruiting tools.

comment by XiXiDu · 2011-12-14T09:57:55.515Z · LW(p) · GW(p)

What about The Lifespan Dilemma and Pascal's Mugging?

Should we penalize computations with large space and time requirements? This is a hack that solves the problem, but is it true? Are computationally costly explanations less likely? Should I think the universe is probably a coarse-grained simulation of my mind rather than real quantum physics, because a coarse-grained human mind is exponentially cheaper than real quantum physics? Should I think the galaxies are tiny lights on a painted backdrop, because that Turing machine would require less space to compute?

Given that, in general, a Turing machine can increase in utility vastly faster than it increases in complexity, how should an Occam-abiding mind avoid being dominated by tiny probabilities of vast utilities?

It seems that as long as you don't solve those problems a rational agent might have a nearly infinite incentive to expend all available resources on attempting to leave this universe, hack the matrix or undertake other crazily seeming stunts.

Replies from: jacob_cannell, timtyler
comment by jacob_cannell · 2011-12-14T21:45:36.741Z · LW(p) · GW(p)

What about The Lifespan Dilemma and Pascal's Mugging?

These are really only problems for agents with unbounded utility functions. This is a great example of over-theorizing without considering practical computational limitations. If your AI design requires double (or even much higher) precision arithmetic just to evaluate it's internal utility functions, you have probably already failed.

Consider the extreme example of bounded utility functions: 1-bit utilities. A 1-bit utility function can only categorize futures into two possible shades: good or bad. This by itself is not a crippling limitation if the AI considers a large number of potential futures and computes a final probability-weighted decision score with higher precision. For example when considering two action paths A and B, a monte carlo design could evaluate out a couple hundred futures branching from A and B, assign each a 0 or 1, and then add them up into a tally requiring precision proportional to the number of futures evaluated (in this case, around 8-bit).

This extremely bounded design would need to do far more future-simulation to compensate for it's extremely low-granularity utility rankings: for example when playing chess, it could only categorize board states as 'likely win' or 'likely loss'. Thus it would need to have a higher ply-depth than algos that use higher-bit depth evaluations. But even so, this only amounts to a performance efficiency disadvantage, not a fundamental limitation.

If we extrapolate a 1-bit friendly AI to planning humanity's future, it would collapse all futures that humanity found largely 'desirable' into the same score of 1, with everything else being 0. If it's utility classifier and future modelling is powerful enough this design can still work.

And curiously a 1-bit utility function gives more intuitively reasonable results in the Lifespan Dilemma or Pascal's Mugging. Assuming dying in an hour is a 0-utility outcome and living for at least a billion years is a 1, it would never take any wagers increasing it's probability of death. And it would be just as un-susceptible to Pascal Mugging.

Just to be clear, I'm not really advocating simplistic 1-bit utilities. What is clear is that human's internal utility evaluations are bounded. This probably comes from practical computational limitations, but likely future AI's will also have practical limitations.

Replies from: Nornagest
comment by Nornagest · 2011-12-14T22:36:43.677Z · LW(p) · GW(p)

Bounded utilities -- especially strongly bounded ones like your 1-bit probability-weighted utility function -- give you outcomes that depend crucially on the probability of a world-state's human-relative improvement versus the probability of degeneration. Once a maximal state has been reached, the agent has an incentive to further improve it if and only if that makes the maintenance of the state more likely. That's not really a bad outcome if we've chosen our utility terms well (i.e. not foolishly ignored the hedonic treadmill or something), but it's substantially less awesome than it could be; I suspect that after a certain point, probably easily achievable by a superintelligence, the probability mass would shift from favoring a development to a maintenance mode.

The first thing that comes to mind is a scenario like setting up Earth as a nature preserve and eating the rest of the galaxy for backup feedstock and as insurance against astronomical-level destructive events. That's an unlikely outcome -- I'm already thinking of ways it could be improved upon -- but it ought to serve to illustrate the general point.

Replies from: jacob_cannell
comment by jacob_cannell · 2011-12-14T23:42:10.219Z · LW(p) · GW(p)

Once a maximal state has been reached, the agent has an incentive to further improve it if and only if that makes the maintenance of the state more likely.

This is true, but much depends on what is considered a 'maximal state'. If our 1-bit utility superintelligence predicts future paths all the way to the possible end states of the universe, then it isn't necessarily susceptible to getting stuck in maintenance states along the way. It all depends on what sub-set of future paths we classify as 'good'.

Also keep in mind that the 1-bit utility model still rates entire future paths, not just end future states. So let's say for example that we are really picky and we only want Tipler Omega Point end-states. If that is all we specify, then the SuperIntelligence may take us through a path that involves killing off most of humanity. However, we can avoid that by adding further constraints on the entire path: assigning 1 to future paths that end in the Omega Point but also satisfy some arbitrary list of constraints along the way. Again this is probably not the best type of utility model, but the weakness of 1-bit bounded utility is not that it tends to get stuck in maintenance mode for all utility models.

The failure in 1-bit utility is more in the specificity vs feasibility tradeoff. If we make the utility model very narrow and it turns out that the paths we want are unattainable, then the superintelligence will gleefully gamble everything and risks losing the future. For example the SI which only seeks specific Omega Point futures may eat the moon, think for a bit, and determine that even in the best actions sequences, it only has a 10^-99 of winning (according to it's narrow OP criteria). In this case it won't 'fall back' to some other more realistic but still quite awesome outcome, no it will still proceed to transform the universe in an attempt to achieve the OP, no matter how impossible. Unless of course there is some feedback mechanism with humans and utility model updating, but that amounts to circumventing the 1-bit utility idea.

comment by timtyler · 2011-12-14T15:09:11.959Z · LW(p) · GW(p)

It seems that as long as you don't solve those problems a rational agent might have a nearly infinite incentive to expend all available resources on attempting to leave this universe, hack the matrix or undertake other crazily seeming stunts.

I don't think this is a significant practical problem.

We have built lots of narrow intelligences. They work fine and this just doesn't seem to be much of an issue.

comment by [deleted] · 2011-12-13T15:46:29.965Z · LW(p) · GW(p)

What is the ideal approximation of perfect Bayesianism?

Given that "the ideal" means "the best conceivable", isn't the "ideal approximation of perfect Bayesianism" just perfect Bayesianism?

Replies from: lukeprog
comment by lukeprog · 2011-12-13T19:05:40.365Z · LW(p) · GW(p)

I meant "ideal computable." Fixed.

comment by [deleted] · 2011-12-16T07:06:12.952Z · LW(p) · GW(p)

The link for the 5-and-10 problem is wrong. This perhaps?

comment by lukeprog · 2011-12-14T07:24:29.758Z · LW(p) · GW(p)

I'll keep this document updated on my own site, but I want to make the question titles expand and collapse into multi-paragraph explanations upon clicking on them, something like this. (The question titles will no longer be list elements.) If someone is willing to help me with the Javascript and JQuery, please contact me at luke [at] singularity.org.

Replies from: timtyler
comment by timtyler · 2011-12-14T11:01:51.514Z · LW(p) · GW(p)

My own tool for doing that with is here: http://treeeditor.com/tree_viewer/

Replies from: lukeprog
comment by lukeprog · 2011-12-27T22:11:02.296Z · LW(p) · GW(p)

Thanks. A volunteer from Facebook wrote me up some custom JQuery stuff.

comment by Matt_Simpson · 2011-12-14T02:57:50.483Z · LW(p) · GW(p)

How would an ideal agent with infinite computing power choose an ideal prior? (A guess: we'd need an anthropic, non-Cartesian, higher-order-logic version of Solomonoff induction.) How can this be process be approximated computably and tractably?

A question for those who know such things. What's the issue with Solomonoff induction here? Is it that the Solomonoff prior doesn't take into account certain prior information that we do have, but isn't based on simply updating from the original (Solomonoff) prior?

Replies from: timtyler, PeterKinnon
comment by timtyler · 2011-12-14T17:27:27.972Z · LW(p) · GW(p)

What's the issue with Solomonoff induction here?

"Higher-order-logic": reputedly down to concerns about uncomputability - which don't seem very interesting to me.

"Anthropic: I figure that can be dealt with in the same way as any other reference machine problem: by "conditioning" it by exposing it to the world.

"Cartesian": I think that's probably to do with this (from E.Y.):

AIXI devises a Cartesian model of a separated environmental theatre, rather than devising a model of a naturalistic reality that includes AIXI.

Fun stuff - but nothing specifically to do with Solomonoff induction. The papers on Orseau's Mortal Universal Agents page address this issue.

comment by PeterKinnon · 2011-12-16T21:24:42.822Z · LW(p) · GW(p)

Unfortunately this entire discussion is deeply flawed.

Why? GIGO - Garbage in - Garbage out.

However good the logical systems used for processing information they are of no avail without meaningful input data.

Present technologies cannot be used as a basis for prediction because of the unexpected bifurcations and inherent non-linearities in technological developments.

Further problems stem from the use of the very inappropriate buzz-word "Singularity". Certainly a dramatic change is imminent, but this is better considered as a phase transition - the emergence of a new and dominant non-biological phase of the on-going evolutionary process that can be traced back at least as far as stellar nucleosynthesis.

Indeed, the inevitable self-assembly of this new entity can be clearly observed in what we at present call the Internet.

The broad evolutionary model which supports this proposition is outlined (very informally) in my latest book: "The Goldilocks Effect: What Has Serendipity Ever Done For Us?" It is a free download in e-book formats from the "Unusual Perspectives" website