Comment by Heighn on Top 9+2 myths about AI risk · 2016-04-12T17:04:58.175Z · LW · GW

Although, now that I think about it, this survey is about risks before 2100, so the 5% risk of superintelligent AI might be that low because some of the responders belief such AI not to happen before 2100. Still, it seems in sharp contrast with Yudkowsky's estimate.

Comment by Heighn on Top 9+2 myths about AI risk · 2016-04-12T16:56:16.375Z · LW · GW

Agreed, especially when compared to

Comment by Heighn on Top 9+2 myths about AI risk · 2016-04-12T12:49:27.987Z · LW · GW

Commenting on the first myth, Yudkowsky himself seems to be pretty sure of this when reading his comment here: I know Yudkowsky's post is written after this LessWrong article, but it still seems relevant to mention.

Comment by Heighn on A pessimistic view of quantum immortality · 2014-03-06T18:22:32.905Z · LW · GW

By the same logic of Quantum Immortality, shouldn't we expect never to fall asleep, since we can't observe ourselves while being asleep?

Comment by Heighn on If MWI is correct, should we expect to experience Quantum Torment? · 2014-03-05T19:11:45.114Z · LW · GW

I was thinking about this post and thought up the following experiment. Suppose, by some quantum mechanism, Bob has a 50% probability of falling asleep for the next 8 hours and a 50% probability of staying awake for the next 8 hours. By the same logic as QI, should Bob expect (with 100% certainty) to be awake after 2 hours, since he cannot observe himself being asleep? I would say no. But then, doesn't QI fail as a result?