A possible solution to pascals mugging.

post by staticIP · 2012-10-13T00:00:38.843Z · LW · GW · Legacy · 57 comments

Contents

57 comments

Now I tend not to follow this form very much, so please excuse me if this has been suggested before. Still, I don't know that there's anyone else on this board who could actually carry out these threats.

 

If anyone accepts a pascals mugging style trade off with full knowledge of the problem, then I will slowly torture to death 3^^^^3 sentient minds. Or a suitably higher number if they include a higher (plausible, from my external viewpoint) number. Rest assured I can at least match their raw computing power from where I am. Good luck.

 

EDIT: I'm told that Eleizer proposed a similar solution over here, although more eloquently then I have.

57 comments

Comments sorted by top scores.

comment by [deleted] · 2012-10-13T00:19:13.522Z · LW(p) · GW(p)

That's not a solution, it's just another mugging.

What's your point?

A solution to pascals mugging answers the question: What do we do when there is a very tiny (O(2^-x)) chance that some random thing has incredibly huge (O(3^^^3)) importance?

Replies from: staticIP
comment by staticIP · 2012-10-13T00:32:12.726Z · LW(p) · GW(p)

The obvious solution is that I, as an incredibly powerful omnipotent god being, will mug anyone who spends too many resources dealing with pascals mugging style issues. If you have sufficient evidence that your particular example outweighs the risk of me doing that, then it's not exactly a pascals mugging, is it?

At the very least it shows there's a point behind which it's no longer useful to consider a very small probability in your utility calculation. That point is where the probability of your hypothesis being true is close to being as probable as my claim being true. So now the big question is "how unlikely is my claim"?

Replies from: ewang
comment by ewang · 2012-10-13T02:18:38.907Z · LW(p) · GW(p)

What do we do when there is a very tiny (2^-x) chance that THIS random thing has incredibly huge (O(3^^^^3)) importance?

Replies from: staticIP
comment by staticIP · 2012-10-13T02:27:49.212Z · LW(p) · GW(p)

Well it depends. Is the likelihood of it affecting reality around the same likelihood as me torturing everybody with my god powers? If it's significantly higher then that, I'd argue it's not an actual pascals mugging. Hard to know exactly where to draw the line though.

If it's not significantly higher then that, then you have to balance against my threat to torture everybody.

comment by gwern · 2012-10-13T00:44:06.951Z · LW(p) · GW(p)

If anyone accepts a pascals mugging style trade off with full knowledge of the problem, then I will slowly torture to death 3^^^^3 sentient minds. Or a suitably higher number if they include a higher (plausible, from my external viewpoint) number. Rest assured I can at least match their raw computing power from where I am.

I go for a walk and meet a mugger. He hears about your precommitment and says, 'ah well, I know what reward I can offer you which both overcomes your expressed probability that I will pay of 1/3^^^^3 and also the threat of -3^^^^3: I will offer you not 3^^^^3 but 3^^^^^^3! If you will simply multiply out the expected value and then subtract staticIP's threat, you will find this is a very lucrative deal for you!'

I conclude his logic is unimpeachable and the net expected value so vast I would be a fool to not give him $5, and promptly do so.

Replies from: staticIP
comment by staticIP · 2012-10-13T01:00:44.479Z · LW(p) · GW(p)

a suitably higher number if they include a higher (plausible, from my external viewpoint) number. Rest assured I can at least match their raw computing power from where I am.

So I'll add a couple of orders of magnitude. I'm certain he will as well. It ultimately comes down to which of us you believe more. Do you have any reason to believe him more then myself?

Replies from: gwern
comment by gwern · 2012-10-13T01:46:10.156Z · LW(p) · GW(p)

So I'll add a couple of orders of magnitude. I'm certain he will as well.

So then your solution doesn't work. Any precommitment you make a mugger can just trivially overcome.

It ultimately comes down to which of us you believe more.

You're not following the process here. 'Which of you I believe more' is already being compensated for by additional increases in the promised reward.

Replies from: staticIP
comment by staticIP · 2012-10-13T02:00:17.346Z · LW(p) · GW(p)

Any precommitment you make a mugger can just trivially overcome.

The same is true for me. Any threat the mugger makes I can trivially overcome, being a god entity and all that.

I'll reward/punish a suitably higher number of negative or positive utilitons if they include a higher number of negative or positive utilitons. Rest assured I can at least match their raw computing power from where I am.

I'm always going to claim that my threat is equal to or greater then their threat. Make sense?

Replies from: gwern
comment by gwern · 2012-10-13T02:08:16.238Z · LW(p) · GW(p)

I'm always going to claim that my threat is equal to or greater then their threat. Make sense?

And their claim afterwards? Any threat you can make, they can make. You see why this is a dead end?

(And what kind of decision theory requires a third party to make precommitments before you can make the right decision, anyway?)

Replies from: BerryPick6, staticIP
comment by BerryPick6 · 2012-10-13T02:11:33.103Z · LW(p) · GW(p)

(And what kind of decision theory requires a third party to make precommitments before you can make the right decision, anyway?)

3PDT, perhaps? ;)

comment by staticIP · 2012-10-13T02:21:14.531Z · LW(p) · GW(p)

And their claim afterwards? Any threat you can make, they can make. You see why this is a dead end?

That's kind of what I'm trying to point out here. It is a dead end, but I'm actually claiming the below. Sure, someone else can also claim the below as well. We can both make the claims. Now, who do you believe more?

But lets formalize my claim. My claim is that I will make n+1 utilitons happen if n is positive or n-1 utilitons happen is n is negative, as long as you do the opposite of what they tell you to do.

Where n is how many utilitons they offer given any result.

I'm outside of your conception of time. So if they make the threat after this is of no concern to me.

Replies from: gwern
comment by gwern · 2012-10-13T02:29:02.169Z · LW(p) · GW(p)

I'm outside of your conception of time. So if they make the threat after this is of no concern to me.

You can't just wave your hands like that. What if the mugger offers a more complicated deal like a 2-step reward, where the second step overcomes your penalty? Are you just going to say 'fine my precommitment is to the net value'? But then the mugger can just toss in a Turing machine to his offer, and now your attempt to analyze his offer is equivalent to solving a halting problem! If you claim to have an oracle on hand, so can he, and that just relativizes the problem because with an oracle, now there are meta-halting problems... etc.

Your strategy doesn't work. Deal with it.

Replies from: staticIP
comment by staticIP · 2012-10-13T02:37:56.380Z · LW(p) · GW(p)

Your strategy doesn't work. Deal with it.

I'm afraid you still haven't shown me enough evidence. If you'll bear with me a moment longer, I'll try to explain it better.

But then the mugger can just toss in a Turing machine to his offer, and now your attempt to analyze his offer is equivalent to solving a halting problem

I want you to think of exactly the consequences of that. Now you have two omnipotent god entities, each offering you a more or less random result. All you know about the result is that it's going to be positive or negative. One offers you a random positive result if you kill a baby, one offers you a random negative result if you kill a baby. Do you get, on average and over a very large sample size, more utility from killing the baby or not killing her?

Replies from: gwern
comment by gwern · 2012-10-13T03:33:06.733Z · LW(p) · GW(p)

Now you have two omnipotent god entities, each offering you a more or less random result. All you know about the result is that it's going to be positive or negative.

One of which is in a temporally advantaged position in which he can do anything you can do and do more in addition to that - a strictly superior position.

Do you get, on average and over a very large sample size, more utility from killing the baby or not killing her?

Without auxiliary arguments about what sample space we are drawing from, I don't see how you could possibly come to any kind of conclusion about this.

Replies from: staticIP
comment by staticIP · 2012-10-13T12:24:07.531Z · LW(p) · GW(p)

One of which is in a temporally advantaged position in which he can do anything you can do and do more in addition to that - a strictly superior position.

Sorry, explain to me how this hypothetical god-being can exceed my threat consistently? Presuming we are both from the same privileged outside-your-time perspective?

comment by Mitchell_Porter · 2012-10-13T08:44:42.765Z · LW(p) · GW(p)

They say a conservative is a liberal mugged by reality. What do you call a utilitarian mugged by unreality?

comment by RobertLumley · 2012-10-13T04:20:52.606Z · LW(p) · GW(p)

The fact that there are 37 comments on this post (and similar posts) is rapidly convincing me that the new -5 karma penalty for replying to downvoted things is actually a really good idea.

Replies from: army1987, staticIP
comment by A1987dM (army1987) · 2012-10-13T10:30:00.350Z · LW(p) · GW(p)

And the fact that many of those comments are on Top Comments Today shows that lots of people likely disagree with you.

Replies from: RobertLumley
comment by RobertLumley · 2012-10-13T15:58:56.971Z · LW(p) · GW(p)

No, it shows that a lot of people upvote comments disagreeing with this post. Downvoted threads usually get a lot of upvotes for commenters explaining what the OP is doing wrong. It helps signal to the OP that they are wrong and that the commenter's explanation is correct. The sad side effect, which Eliezer is so opposed to is that these upvotes encourage feeding the trolls. Up until now, I was thoroughly unconvinced that this was a problem, but this thread and this one are convincing me otherwise, which was the point of my comment.

As a side point, the grandparent is on the top comments today list too... So if your criteria for "people agree with and support this comment" is "is on Top Comments Today" then you've run into a paradox.

Tangentially, some of those comments are at 0 karma, and are still on that list. This doesn't make much sense to me, but honestly I've never looked at that feature before, and probably will continue not to.

Replies from: army1987, army1987
comment by A1987dM (army1987) · 2012-10-14T09:18:02.355Z · LW(p) · GW(p)

Tangentially, some of those comments are at 0 karma, and are still on that list. This doesn't make much sense to me, but honestly I've never looked at that feature before, and probably will continue not to.

I had only looked at the first couple screenfuls of the list when replying to you, and most comments were from this thread and upvoted quite a bit. Anyway, the thing does occasionally misplace comments (e.g. heavily downvoted comments mysteriously in the middle of it), though I think that when there are lots of zero-karma comments at the bottom there really aren't any more upvoted comments from the last 24 hours.

comment by A1987dM (army1987) · 2012-10-14T09:13:06.564Z · LW(p) · GW(p)

If people thought it's bad to feed the trolls, even when disagreeing with them, they'd downvote all replies to trolls.

comment by staticIP · 2012-10-13T12:35:00.143Z · LW(p) · GW(p)

I understand that a lot of issues are solved, like the existence of god and so on, but I for one still haven't gotten an appropriate explanation as to why my claim, which seems perfectly valid to me, is incorrect. That proposal is going to further hinder this kind of discussion and debate.

And as far as I can tell, I'm correct. It's honestly very concerning to me that a bunch of lesswrongers have failed to follow this line of reasoning to its natural conclusion. Maybe I'm just not using the correct community-specific shibboleths, but the only one who's actually followed through on the logic is gwern. I look forward to seeing his counter reply to this.

Replies from: buybuydandavis
comment by buybuydandavis · 2012-10-14T02:26:04.269Z · LW(p) · GW(p)

I think you're getting at an important problem, and have taken a step toward the solution.

How do we deal with choice in the face of fundamentally arbitrary assertions?

One way, at least, is to see if there is another equivalently arbitrary assertion that would lead you to make the opposite choice.

comment by roryokane · 2012-10-13T00:40:02.825Z · LW(p) · GW(p)

This solution doesn’t work. Why? Because I pledge that if anyone fails to accept a “Pascal’s Mugging style trade-off with full knowledge of the problem, then I will slowly torture to death 3^^^^3 sentient minds”. I’ve just canceled out your pledge.

You could say all your allies take the same pledge as you, and you have more allies than me, but that’s getting too far into the practicalities of our lives and too far away from a general solution. A general solution can’t assume that the person considering whether to accept a Mugging will have heard either of our pledges, so the person would be unable to take those pledges into account for their decision.

I don’t know the actual solution to Pascal’s Mugging myself. I’ve pasted my outline-form notes on it so far into a reply to this comment, in case they’re useful.

Replies from: Epiphany, Vladimir_Nesov, roryokane, staticIP
comment by Epiphany · 2012-10-13T01:51:15.964Z · LW(p) · GW(p)

You're not thinking big enough. If anyone ever accepts a Pascal's mugging again, my fuzzy celery God will execute a worse Pascal's mugging than any other in existence no matter what the original Pascal's mugging is.

P.S. You'll never find out what muggings fuzzy celery God executes because they're always unpredictable. This makes them impossible to disprove.

(My brother always hated having fights like this with me.)

Replies from: staticIP
comment by staticIP · 2012-10-13T02:03:47.102Z · LW(p) · GW(p)

If we have two gods, one claiming that if I do X, they'll mug me, and one claiming that if I don't do X they'll mug me, well I'm probably going to believe the god that isn't fuzzy and celery...

Replies from: Epiphany
comment by Epiphany · 2012-10-13T02:28:16.681Z · LW(p) · GW(p)

Well that's making the wrong choice, buddy. Other Gods are useless against fuzzy celery God because fuzzy celery God can transform itself at will into the Most Believable God. Don't think of fuzzy celery God as a piece of fuzzy celery. Fuzzy celery God is nothing like that. If an old wise man is the most compelling God-form for you, fuzzy celery God looks like an old wise man. If benevolent Gods are more credible to you, fuzzy celery God becomes benevolent. No matter what Pascal's mugging the person wants to accept, fuzzy celery God will always take on the appearance and traits of the most believable God that the person can conceive of.

comment by Vladimir_Nesov · 2012-10-13T14:51:04.574Z · LW(p) · GW(p)

This solution doesn’t work. Why? Because I pledge that if anyone fails to accept a “Pascal’s Mugging style trade-off with full knowledge of the problem, then I will slowly torture to death 3^^^^3 sentient minds”. I’ve just canceled out your pledge.

Your argument doesn't address the problem with Static_IP's post, and indeed it has exactly the same problem: it is not an argument/explanation/clarification, but instead it's one more mugging, see nyan_sandwich's comment. The problem is not that someone has put out a Pascal's mugging and now we have to pay up, unless the mugger is neutralized in some way. If it turns out that we in fact should pay up, the correct decision is easily performed.

The problem is that this situation is not understood. The theoretical model of expected utility plus some considerations about prior suggest that the correct decision is to pay the mugger, yet other considerations suggest otherwise, and there are potential flaws with the original argument, which motivates a search for better understanding of the situation. Modifying the situation in a way that makes the problem go away doesn't solve the original problem, it instead shifts attention away from it.

comment by roryokane · 2012-10-13T00:42:50.282Z · LW(p) · GW(p)

My unfinished outline-form notes on solving Pascal’s Mugging:

  • Pascal’s Mugger (http://www.nickbostrom.com/papers/pascal.pdf) possible solutions
    • Off – Pascal’s estimate might be farther off than the offered benefit, and how does he know how far to compensate?
    • Counter – there is a (smaller) probability that the man will give you the same amount of Utility only if you refuse. (Also a probability that will give way more Utility if refuse, but probably countered by probability that will give way more if accept.)
    • Known – the gambit is known, so that makes it more likely that he is tricking you – but sadly, no effect, I think.
    • Impossible – [My Dad]’s suspect argument: there is absolutely zero probability of the mugger giving you what he promises. There is no way to both extend someone’s lifespan and make them happy during it.
      • He could just take you out of the Matrix into a place where any obstacles to lengthy happiness are removed. There's still a probability of that, right?
    • God – maybe level of probability involved is similar to that of God's existence, with infinite heaven affecting the decision
    • Long-term – maybe what we should do in a one-shot event is different from what we should do if we repeated that event many times.
    • Assumption – one of the stated assumptions, such as utilitarianism or risk-neutrality, is incorrect and should not actually be held.
comment by staticIP · 2012-10-13T00:51:48.909Z · LW(p) · GW(p)

As an omnipotent god entity I pledge to counter any any attempt at pascals muggings, as long as the mugger actually has the power to do what they say.

I’ve just canceled out your pledge.

Yep. You did, or you would have if you could actually carry through on your threats. I maintain that you can't. Now it's a question of which of our claims is more likely to be true. That's kind of the point here. When you're dealing with that small or a probability then the calculation becomes useless and marred by noise.

If I'm correct, and I'm one of the very few entities capable of doing this, who happen across your world anyway, then I can cancel out your claim and a bunch of future claims. If you're correct then I can't. So the question is, how unlikely are my claims? How unlikely are yours? Are your claims significantly more likely (on the tiny scales we're working with) then mine?

But yes, now that I look at it more in depth (thank you for the links), it's obvious that this is a reiteration of the "counter" solution, but with actual specific and viable threats behind it.

comment by NancyLebovitz · 2012-10-14T13:08:32.321Z · LW(p) · GW(p)

If anyone accepts a pascals mugging style trade off with full knowledge of the problem, then I will slowly torture to death 3^^^^3 sentient minds.

I realize Pascal's Muggings involve highly implausible threats, but I think people are focusing on the 3^^^^3 sentient minds, and I have the more doubts about the surveillance problems. I can sort of imagine the feasibility of torturing 3^^^^3 minds (though I think the canonical form requires infinite torture rather than torturing to death) through some combination of bubbling off universes and automation, but I can't imagine any way of keeping track of whether anyone has accepted a Pascal's Mugging with full knowledge of the problem.

comment by wedrifid · 2012-10-13T00:30:37.493Z · LW(p) · GW(p)

Now I tend not to follow this form very much, so please excuse me if this has been suggested before.

It actually, has. Within the last week or two.

Replies from: staticIP
comment by staticIP · 2012-10-13T00:36:26.616Z · LW(p) · GW(p)

Apparently my search-fu is weak. Would you care to link, or suggest search terms that would make finding it less arduous?

EDIT: found it, I think, over here. One of the obvious issues is that it's not a credible threat.

So far I haven't seen a counter argument over there that satisfies me. If there is anywhere that they go into it more in depth, please do give it a link here.

Replies from: Epiphany
comment by Epiphany · 2012-10-13T01:44:03.704Z · LW(p) · GW(p)

Why not put all the search terms you tried during your search-fu session into a comment. That will make this less likely to happen in the future.

Replies from: staticIP
comment by staticIP · 2012-10-13T01:54:23.591Z · LW(p) · GW(p)

Because I doubt I can remember all of them. Also, I'm not entirely clear on why you have "quotes" around search-fu. It's a pretty accepted term on the internet. Search-fu is the skill one is employing when searching for something.

A more reasonable question seems to me like asking how I arrived at the answer, not asking how I failed to arrive at the answer. I find it odd that you'd go that particular route, and would be very interested if you could expand on why you wanted to see all of my failed attempts, instead of my one successful attempt?

Seeing all my failed attempts is more difficult for me to write out (there were a few of them) and contains less usable information. I'm curios as to why you'd pick that instead of what seems to me like the more reasonable "post the search term that got you the answer".

Replies from: Epiphany
comment by Epiphany · 2012-10-13T02:04:57.184Z · LW(p) · GW(p)

would be very interested if you could expand on why you wanted to see all of my failed attempts, instead of my one successful attempt?

Because all of the other people who want to start threads on solutions to Pascal's muggings will be more likely to find your thread and go "Oh, someone already did this." That will save us from similar "solutions to Pascal's mugging threads" in the future. The worse your search terms failed, the better - that way Pascal's mugging solvers with low search-fu skill (no quotes this time) will be likely to find the existing solutions.

P.S. It's better to put them into complete sentences to avoid search engine penalties.

Replies from: Epiphany, staticIP
comment by Epiphany · 2012-10-13T02:06:21.982Z · LW(p) · GW(p)

My fu cents:

Solutions to Pascal's Muggings. I Solved Pascal's Mugging. Would this Solve Pascal's Mugging? Possible Solution to Pascal's Muggings. Solve for Pascal's Muggings. Pascal's Mugging has been solved. Pascal's Mugging has a solution.

comment by staticIP · 2012-10-13T02:12:48.521Z · LW(p) · GW(p)

Ahh, makes sense. I actually found many different and interesting solutions to pascals mugging with my search terms though. Just not this "counter" solution.

This thread already shows up pretty close to the top for searches of "pascals mugging solutions" that I've attempted. For that exact phrase it's number 3, and has been before you posted this. I don't know that this particular solution needs to be more closely associated with the search terms then it already is.

Replies from: Epiphany
comment by Epiphany · 2012-10-13T02:17:03.947Z · LW(p) · GW(p)

Eliezer presented a counter solution similar to this on Overcoming Bias. search fu

Click here for Flimple Utility!!!

Replies from: staticIP
comment by staticIP · 2012-10-13T02:24:12.884Z · LW(p) · GW(p)

Well done. Roryokane mentioned it up here however.

Replies from: Epiphany
comment by Epiphany · 2012-10-13T02:33:27.112Z · LW(p) · GW(p)

I wonder what Eliezer would DO if he actually got $5 for flimple utility. I kinda want to try sending him five bucks as a psychology experiment.

Replies from: MileyCyrus
comment by MileyCyrus · 2012-10-13T12:48:03.833Z · LW(p) · GW(p)

Done. It's not really a sacrifice since I already donate to the SIAI. I will miss that 5 karma though.

Edit: Apparently you only get the -5 karma if you directly reply to a downvoted comment/post. Replying to reply is fine.

Replies from: Epiphany
comment by Epiphany · 2012-10-13T17:06:47.815Z · LW(p) · GW(p)

What? I don't understand. Did you give him 5$ or 5 karma or what?

Replies from: MileyCyrus
comment by MileyCyrus · 2012-10-13T22:02:44.221Z · LW(p) · GW(p)

Yeah I gave him $5.

Replies from: Epiphany
comment by Epiphany · 2012-10-14T00:07:34.232Z · LW(p) · GW(p)

Did you say it was for flimple utility?

Replies from: MileyCyrus
comment by MileyCyrus · 2012-10-14T02:50:09.217Z · LW(p) · GW(p)

No.

Replies from: Epiphany
comment by Epiphany · 2012-10-14T02:52:59.140Z · LW(p) · GW(p)

Okay, well, I guess he doesn't know.

comment by BerryPick6 · 2012-10-13T02:08:29.216Z · LW(p) · GW(p)

I don't see why there has to be a solution to Pascal's Mugging. Is it so implausible to think that it's just a valid quirk in the system?

Replies from: The_Duck, staticIP
comment by The_Duck · 2012-10-13T03:43:39.574Z · LW(p) · GW(p)

Does this mean you'll pay up if you get Pascal-mugged, because it's a "valid quirk in the system"? If so, I have an amazing offer for you.

If you won't pay up, why not?

comment by staticIP · 2012-10-13T02:14:16.347Z · LW(p) · GW(p)

Err, yes. Maybe it is. That's what I'm trying to find out...

Are you saying that I should take some action with the knowledge that it might just be a quirk in the system? Like not posting my hypothesis?

Replies from: BerryPick6
comment by BerryPick6 · 2012-10-13T02:18:29.130Z · LW(p) · GW(p)

Are you saying that I should take some action with the knowledge that it might just be a quirk in the system? Like not posting my hypothesis?

Not at all. I was probably being unclear, I apologize.

Replies from: staticIP
comment by staticIP · 2012-10-13T02:28:30.501Z · LW(p) · GW(p)

Fair enough. Would you mind explaining your intent then?

comment by DuncanS · 2012-10-13T10:30:19.955Z · LW(p) · GW(p)

If anyone accepts a pascals mugging style trade off with full knowledge of the problem,

Well, it's very well known that Pascal himself accepted it, and I'm sure there are others. So, off you go and do whatever it is you wanted to do.

To be honest, your ability to come through on this threat is a classic example of the genre - it's very, very unlikely that you are able to do it, but obviously the consequences if you were able to would be, er, quite bad. In this case my judgement of the probabilities is that we are completely justified in ignoring the threat.

Replies from: staticIP
comment by staticIP · 2012-10-13T12:27:13.868Z · LW(p) · GW(p)

In this case my judgement of the probabilities is that we are completely justified in ignoring the threat.

Do you consider my pascals mugging to be less likely then the general examples of the genre, or do you think that all pascals muggings" probabilities are that we are completely justified in ignoring the threat."

Replies from: DuncanS
comment by DuncanS · 2012-10-13T21:52:36.667Z · LW(p) · GW(p)

It surely depends on one's estimate of the numbers. It seems worthwhile doing something about possible asteroid impacts, for example.

comment by Epiphany · 2012-10-13T01:39:20.602Z · LW(p) · GW(p)

The fact that we're not dying off right now proves this untrue.

Replies from: staticIP
comment by staticIP · 2012-10-13T01:41:33.018Z · LW(p) · GW(p)

I never threatened to harm you. Yes, on average, you're significantly more likely to be in the torture group then where you are now, but anthropic principle and all that.