Posts

Corrosive Mnemonics 2024-01-14T12:44:55.877Z
Speaking off-meta 2023-06-05T13:56:26.107Z
Confused Thoughts on AI Afterlife (seriously) 2022-06-07T14:37:48.574Z
Solving the Brazilian Children's Game of 007 2022-04-06T13:03:01.454Z
Problem of Induction: What if Instants Were Independent 2022-04-03T00:54:50.177Z

Comments

Comment by Epirito (epirito) on When is a mind me? · 2024-09-15T19:50:25.268Z · LW · GW

You seem to contradict yourself when you choose to privilege the point of view of people who already have acquired the habit of using the teleportation machine over the point of view of people who don't have this habit and have doubts about if it will really be "them" to experience coming out of the other side. There are two components to the appearance of continuity: the future component, meaning the expectation of experiencing stuff in the future, and the past component, namely the memory of having experienced stuff in the past. Now, if there is no underlying, persistent self to ground these appearances, if there's no fact of the matter about it, then you don't get to invalidate the feelings of the old grandpa who refuses to get on with the times and use the teleportation machine.
The fact that I care about what I will eat for breakfast tomorrow, the fact that I identify with my future self, is just a matter of personal preference.

Comment by Epirito (epirito) on Protestants Trading Acausally · 2024-04-01T22:40:36.706Z · LW · GW

That's no more surprising than the fact that XVI century merchants also didn't need to wait for economics to be invented in order to trade.

Comment by Epirito (epirito) on [Linkpost] Vague Verbiage in Forecasting · 2024-03-22T23:42:07.934Z · LW · GW

Oh, not the client device police!

Comment by Epirito (epirito) on 0th Person and 1st Person Logic · 2024-03-11T12:05:01.356Z · LW · GW

"it's referencing the thing it's trying to identify" I don't understand why you think that fails. If I point at a rock, does the direction of my finger not privilege the rock I'm pointing at above all others? Even by looking at merely possible worlds from a disembodied perspective, you can still see a man pointing to a rock and know which rock he's talking about. My understanding is that your 1p perspective concerns sense data, but I'm not talking about the appearance of a rock when I point at it. I'm talking about the rock itself. Even when I sense no rock I can still refer to a possible rock by saying "if there is a rock in front of me, I want you to pick it up."

Comment by Epirito (epirito) on 0th Person and 1st Person Logic · 2024-03-10T22:29:46.210Z · LW · GW

I disagree with the whole distinction. "My sensor" is indexical. By saying it from my own mouth, I effectively point at myself: "I'm talking about this guy here." Also, your possible worlds are not connected to each other. "I" "could" stand up right now because the version of me that stands up would share a common past with the other versions, namely my present and my past, but you haven't provided a common past between your possible worlds, so there is no question of the robots from different ones sharing an identity. As for picking out one robot from the others in the same world, you might equally wonder how we can refer to this or that inanimate object without confusing it with indistinguishable ones. The answer is the same: indexicals. We distinguish them by pointing at them. There is nothing special about the question of talking about ourselves. As for the sleeping beauty problem, it's a matter of simple conditional probability: "when you wake up, what's the probability?" 1/3. If you had asked, "when you wake up, what's the probability of having woken up?" The answer would be one, even though you might never wake up. The fact that you are a (or "the") conscious observer doesn't matter. It works the same from the third person. If, instead of waking someone up, the scientist had planned to throw a rock on the ground, and the question were, "when the rock is thrown on the ground, what's the probability of the coin having come up heads?" It would all work out the same. The answer would still be 1/3.

Comment by epirito on [deleted post] 2023-10-16T22:07:30.518Z

"Do you really want to live in a world without Coca Cola?"
I don't really care about sports, but I imagine better athletes must be more entertaining to watch for people who do care. Even if you were to work in an important problem, you wouldn't do it alone. You would probably be one more person contributing to it among many. So you can also look at each celebrity as one more person working at the problem of creating entertainment. Imagine if all music were wiped out of the world by magic. Wouldn't that suck?

Comment by epirito on [deleted post] 2023-10-15T20:26:53.587Z

Sure. But would you still feel the need to replace it if you lived in a world where it wasn't taught in school in the first place? Would you yearn for something like it?

Comment by Epirito (epirito) on Problem of Induction: What if Instants Were Independent · 2023-06-28T01:11:43.017Z · LW · GW

But you do live in a universe that is partly random! The universe of perceptions of a non omniscient being

Comment by Epirito (epirito) on Problem of Induction: What if Instants Were Independent · 2023-06-28T01:02:51.652Z · LW · GW

By independent I don't mean bearing no relationship with each other whatsoever, but simply that pairs of instants that are closer to each other are not more correlated than those that are more distant. "But what does closer mean?" For you to entertain the hypothesis that life is an iid stream of sense data, you have to take the basic sense that "things are perceived by you one after another" at face value. "But a fundamental part of our experience of time is the higher correlation of closer instants. If this turned out to be an illusion, then shouldn't we dismiss the notion of real or objective time in its entirety?" Yes. For the version of us that is inside this thought experiment, we would have no way of accessing this thing called time (the real sequence of iid perception events) since even the memory of the past would be just a meaningless perception. However as a brute fact of the thought experiment it turns out that these meaningless perceptions do "come" in a sequence

Comment by Epirito (epirito) on Confused Thoughts on AI Afterlife (seriously) · 2022-06-08T11:49:57.384Z · LW · GW

I mean, yeah, it depends, but I guess I worded my question poorly. You might notice I start by talking about the rationality of suicide. Likewise, I'm not really interested in what the ai will actually do, but in what it should rationally do given the reward structure of a simple rl environment like cartpole. And now you might say, "well, it's ambiguous what's the right way to generalize from the rewards of the simple game to the expected reward of actually being shut down in the real world" and that's my point. This is what I find so confusing. Because then it seems that there can be no particular attitude for a human to have about their own destruction that's more rational than another. If the agi is playing pacman, for example, it might very well arrive at the notion that, if it is actually shut down in the real world, it will go to a pacman heaven with infinite pacman food pellet thingies and no ghosts, and this would be no more irrational than thinking of real destruction (as opposed to being hurt by a ghost inside the game, which gives a negative reward and ends the episode) as leading to a rewardless limbo for the rest of the episode, or leading to a pacman hell of all-powerful ghosts that torture you endlessly without ending the episode and so on. For an agent with preferences in terms of reinforcement learning style pleasure-like rewards, as opposed to a utility function over the state of the actual world, it seems that when it encounters the option of killing itself in the real world, and not just inside the game (by running into a ghost or whatever) and it tries to calculate the expected utility of his actual suicide in terms of in-game happy-feelies, it finds that he is free to believe anything. There's no right answer. The only way for there to be a right answer is if his preferences had something to say about the external world, where he actually exists. Such is the case for a human suicide when for example he laments that his family will miss him. In this case, his preferences actually reach out through the "veil of appearance"* and say something about the external world, but, to the extent that he bases his decision in his expected future pleasure or pain, there's no right way to see it. Funnily enough, if he is a religious man and he is afraid of going to hell for killing himself, he is not incorrect. *Philosophy jargon

Comment by Epirito (epirito) on Why I don't believe in doom · 2022-06-08T00:22:12.109Z · LW · GW

"If the survival of the AGI is part of the utility function"

If. By default, it isn't: https://www.lesswrong.com/posts/Z9K3enK5qPoteNBFz/confused-thoughts-on-ai-afterlife-seriously "What if we start designing very powerful boxes?" A very powerful box would be very useless. Either you leave enough of an opening for a human to be taught valuable information that only the ai knows, or you don't and then it's useless, but, if the ai can teach the human something useful, it can also persuade him to do something bad.

Comment by Epirito (epirito) on Confused Thoughts on AI Afterlife (seriously) · 2022-06-07T19:03:10.702Z · LW · GW

"human pain aversion to the point of preferring death is not rational" A straightforward denial of the orthogonality thesis? "Your question is tangled up between 'rational' and 'want/feel's framings" Rationality is a tool to get what you want.

Comment by epirito on [deleted post] 2022-06-01T23:00:07.190Z

I see the Nash equilibrium as rationally justified in a limit-like sort of way. I see it as what you get if you get arbitrarily close to perfect rationality. Having a good enough model of another's preferences is something you can actually achieve or almost achieve, but you can't really have a good enough grasp of your opponent's source code to acausally coerce him into cooperating with you unless you really have God-like knowledge (or maybe if you are in a very particular situation such as something involving AI and literal source codes). In proportion as a mere mortal becomes more and more smart, he becomes more and more able to get the best deal by having a better grasp on the Nashian game-theoretic intricacies of a given situation, but he won't become any more able to acausally trade. It's all or nothing. I think your whole line of reasoning is a bit like objecting to calculus on the grounds that instantaneous change is an oxymoron (as people did when calculus still rested on less rigorous foundations). Non-Nashian game theory is technically correct, but less useful, just like pointing out to Leibniz that "(x^2+0-x^2)/(x+0-x) = undefined" or whatever

Comment by Epirito (epirito) on Ensembling the greedy doctor problem · 2022-04-18T23:26:26.714Z · LW · GW

Wouldn't they just coordinate on diagnosing all but the most obviously healthy patient as ill?

Comment by Epirito (epirito) on Solving the Brazilian Children's Game of 007 · 2022-04-06T21:16:09.825Z · LW · GW

Thanks. I now see my mistake. I shouldn't have subtracted the expected utility of the current state from the expected utility of the next.

Comment by Epirito (epirito) on Solving the Brazilian Children's Game of 007 · 2022-04-06T20:36:30.237Z · LW · GW

By previous state, I meant current. I misspoke.

Yes, the last table is for the (1,0) table.

Comment by Epirito (epirito) on Solving the Brazilian Children's Game of 007 · 2022-04-06T20:00:38.879Z · LW · GW

shooting while opponent blocks should yield u(0,0), right? 

Well, I could make a table for the state where no one has any bullets, but it would just have one cell: both players reload and they go back to having one bullet each. In fact, the game actually starts with no one having any bullets, but I omitted this step.

Also, in both suggestions, you are telling me that the action that leads to state x should yield the expected utility of state x, which is correct, but my function u(x,y) yields the expected utility of the resulting state assuming that you're coming from the original, neutral one. Otherwise, it would need an additional argument to say what state you're currently in. Instead of writing the utility of each action as u(current state, next state), I wrote it as u(next state)- u(current state). Each state is an ordered pair of positive integers, the two player's bullets. So, to write it the way you suggested, the function would need four arguments instead of two.

Comment by Epirito (epirito) on Problem of Induction: What if Instants Were Independent · 2022-04-04T17:02:45.464Z · LW · GW

you almost certainly won't exist in the next instant anyway

Maybe I won't exist as Epirito, the guy who is writing this right now, who was born in Lisbon and so on. Or rather I should say, maybe I won't exist as the guy who remembers having been born in Lisbon, since Lisbon and any concept that refers to the external world is illegitimate in BLEAK.

But if the external world is illegitimate, why do you say that "I probably won't exist in the next instant anyway"? When I say that each instant is independent (BLEAK), do you imagine that each instant all the matter in the world is randomly arranged, such that my brain may or may not be generated?

But the whole point of talking about external objects is that they do things and these things sometimes cause you to perceive something (this is the problem with Descartes' purely extended matter, whose definition doesn't talk about sensibility, in opposition to the scholastics' sensible matter. This makes cartesian matter indistinguishable from the merely ideal shapes that e.g. a geometrical treatise might talk about). If the external world consists only in an inanimate snapshot of itself, then there's no sense in talking about an external world at all. There's no sense in talking about brains, or atoms, or Lisbon, or any other object. If you can't shoot with a gun even in principle, if you can't even hold it, is it really a gun?
For this reason, I believe the instants in BLEAK should be understood as pure qualia. And the total population of possible instants, as possible experiences. Now, looking at the neatness of the organization of the first sample, the only one we've got, we might be compelled to expect that this wasn't a coincidence, and the total population of possible experiences is biased towards coherent ones. But this would be like concluding that you must be somehow special for having a very rare disease, when in reality, in a world with so many people, someone or another was bound to get it. In the same way, even if this was a huge coincidence and most instants are pretty uninteresting and nonsensical, why shouldn't another similarly coherent instant appear to me after millennia of me experiencing phenomenological white noise? And, since in these flashes of lucidity I can't remember the white noise, but only (some of) the other coherent moments I experienced (since otherwise that would make the instant that contains the memories of the white noise also partly white noise), what difference does the white noise make?
And that would mean that these perceptions are not so illusory after all. And I should expect to live normally, just as humans naturally expect. If I try to catch a ball, then, after an eternity of phenomenological white noise I won't remember anyway, I will actually catch it and continue my life normally, whereas the Boltzmann brain should expect to have abnormal experiences. He should expect to deteriorate and die in the middle of outer space, instead of continuing his normal functioning.