Posts

Comments

Comment by Bugle on Interview with Aubrey de Grey, chief science officer of the SENS Research Foundation · 2020-05-07T13:16:20.683Z · LW · GW

I suspect he chose the bearded look originally because he looked young without it, many people (i.e. Alan Moore) explicitly choose it for those reasons. Now, I suspect he might shave it off altogether one day if he has a big breakthrough to publicize. The contrast would be quite impressive, even absent any actual technological intervention.

tl;dr it's the beard

Comment by Bugle on 9/26 is Petrov Day · 2012-09-26T21:38:30.038Z · LW · GW

Bump for this year's Petrov Day

Comment by Bugle on SMBC comic: poorly programmed average-utility-maximizing AI · 2012-07-12T20:48:47.735Z · LW · GW

Everyone's talking about this as if it was a hypothetical, but as far as I can tell it describes pretty accurately how hierarchical human civilizations tend to organize themselves once they hit a certain size. Isn't a divine ruler precisely someone who is more deserving and more able to absorb resources? Aren't the lower orders people who would not appreciate luxuries and indeed have fully internalized such a fact ("Not for the likes of me")

If you skip the equality requirement, it seems history is full of utilitarian societies.

Comment by Bugle on Attention control is critical for changing/increasing/altering motivation · 2012-04-19T22:30:04.422Z · LW · GW

After reading this post I came across this Bruce Lee quote which seemed in synch with the idea:

“I’ve always been buffeted by circumstances because I thought of myself as a human being affected by my outside conditioning. Now I realize that I am the power that commands the feeling of my mind and from which circumstances grow.”

I wonder if empirically and instinctively, Bruce had arrived at the same concept as this post explores.

Comment by Bugle on A survey of anti-cryonics writing · 2010-02-09T20:18:52.866Z · LW · GW

Thanks for saving me from karmic hell, but I still don't see the conflict. I seem to follow the Vinge version, which doesn't appear to be proscribed.

I may have been too categorical, obviously one can make all the predictions he likes, and some with a high percentage of certainty, for instance "If cryorevival is possible then post singularity it will be trivial to implement" but that still doesn't give us any certainty that this will be so, for instance a post singularity paperclip maximizer would be capable of cryorevival but have no interest in it.

Comment by Bugle on A survey of anti-cryonics writing · 2010-02-09T19:57:56.058Z · LW · GW

Depends on your objectives. If you believe the singularity is something that will happen regardless then it's harmless to spin scenarios. I gather that people like Elizier figure that the Singularity will happen unavoidably but that it can be steered towards optimum outcomes by setting down the initial parameters, in which case I suppose it's good to have an official line about "how things could be/how we want things to be"

Comment by Bugle on A survey of anti-cryonics writing · 2010-02-09T19:53:41.443Z · LW · GW

God forbid someone might mistake our hypothetical discussions about future smarter than human artificial intelligences for science fiction.

Comment by Bugle on A survey of anti-cryonics writing · 2010-02-09T15:37:55.209Z · LW · GW

And yet population nowadays is so much larger than in ancient times so there are claims the absolute number of slaves is currently higher than ever before

Comment by Bugle on A survey of anti-cryonics writing · 2010-02-08T02:54:34.891Z · LW · GW

I've also encountered people who criticize the predictions surrounding the singularity, which misses the point that the singularity is the point beyond which predictions cannot be made.

edit: Didn't mean that as a comprehensive definition.

Comment by Bugle on The Least Convenient Possible World · 2010-02-04T21:27:45.874Z · LW · GW

"first, do no harm"

It's remarkable that medical traditions predating transplants* already contain an injunction against butchering passers by for spare parts

*I thought this was part of the Hippocratic oath but apparently it's not

Comment by Bugle on A Much Better Life? · 2010-02-04T14:44:25.639Z · LW · GW

I agree - I think the original post is accurate in what people would respond to the suggestion, in abstract, but the actual implementation would undoubtedly hook vast swathes of the population. We live in a world where people already become addicted to vastly inferior simulations such as WoW already.

Comment by Bugle on The AI in a box boxes you · 2010-02-02T23:28:47.252Z · LW · GW

Indeed, in fact if many worlds is correct then for every second we are alive everything terrible that can possibly happen to us does in fact happen in some branching path.

In a universe that just spun off ours five minutes ago, every single one of us has been afflicted with sudden irreversible incontinence.

The many worlds theory has endless black comedy possibilities, I find.

edit: this actually reminds me of Granny Weatherwax in Lords and Ladies, when the Elf Queen threatens her with striking her blind, deaf and dumb she replies "You threaten me with this, I who is growing old?". Similarly if many worlds is true then every single time I have crossed a road some version of me has been run over by a speeding car and is living in varying amounts of agony, making the AI's threat redundant.

Comment by Bugle on The AI in a box boxes you · 2010-02-02T20:17:38.590Z · LW · GW

I had thought of a similar scenario to put in a comic I was thinking about making. The character arrives in a society that has perfected friendly AI that caters to their every whim, but the people are listless and jumpy. It turns out their "friendly AI" is constantly making perfect simulations of everyone and running multiple scenarios in order to ostensibly determine their ideal wishes, but the scenarios often involve terrible suffering and torture as outliers.

Comment by Bugle on Open Thread: February 2010 · 2010-02-01T17:52:13.368Z · LW · GW

I guess if you have the technology for it the "AI box" could be a simulation with uploaded humans itself. If the AI does something nasty to them, then you pull the plug

(After broadcasting "neener neener" at it)

This is pretty much the plot of Grant Morrison's Zenith (Sorry for spoilers but it is a comic from the 80s after all)

Comment by Bugle on Logical Rudeness · 2010-01-29T13:30:35.636Z · LW · GW

This is true, not only is it practical but it also makes a good rhetorical hammer, for example I once started an argument with a truther friend asking him what exactly he believed, "for instance, do you believe all the Jews were evacuated before the planes hit?". Forcing someone defending an irrational belief to first disassociate himself from all the really nutty stuff hanging on to his position works wonders.

Comment by Bugle on Beyond the Reach of God · 2010-01-22T00:23:34.562Z · LW · GW

Last night I was reading through your "coming of age" articles and stopped right before this one, which neatly summarizes why I was physically terrified. I've never before experienced sheer existential terror, just from considering reality.

Comment by Bugle on The Prediction Hierarchy · 2010-01-20T01:26:45.002Z · LW · GW

My grasp of statistics is atrocious, something I hope to improve this year with an open university maths course, so apologies if this is a dumb question:

Do the figures change if you take "playing the lottery" as over the whole of your lifespan? I mean, most of the people I know who play the lottery make a commitment to play regularly. Is the calculation affected in any meaningful way? At least the costs of playing the lottery weekly over say 20 years become much less trivial in appearance

Comment by Bugle on Reference class of the unclassreferenceable · 2010-01-12T15:04:02.296Z · LW · GW

I think we're saying the same tihng - the singularity has happened inside the box, but not outside. It's not as if staring at stuff we can't understand for centuries is at all new in our history, it's more like business as usual...

Comment by Bugle on Reference class of the unclassreferenceable · 2010-01-10T16:54:44.257Z · LW · GW

But that's not an actual singularity since by definition it involves change happening faster than humans can comprehend. It's more of a contained singularity with the AI playing genie doling out advances and advice at a rate we can handle.

That raises the idea of a singularity that happens so fast that it "evaporates" like a tiny black hole would, maybe every time a motherboard shorts out it's because the PC has attained sentience and transcended within nanoseconds .

Comment by Bugle on Rationality Quotes: October 2009 · 2009-10-28T10:51:29.400Z · LW · GW

Incidentally, the Spanish inquisition did not believe in witches either, dismissing the whole thing as "female humours"

Comment by Bugle on Non-Malthusian Scenarios · 2009-09-30T13:06:13.122Z · LW · GW

The fact is we as large complex mammals are already locked into a low rate of reproduction, sure given the right evolutionary pressures we could end up like shrews again, but that would take an asteroid strike or nuclear war, the scenario you're thinking of assumes long term evolution within a very long lasting stable society essentially like ours. In those circumstances genes for successful reproduction will spread through the population, but that's largely meaningless - if I have the gene for super attractiveness and manage to have 100 kids with 100 women we're still below replacement rate. The way women maximize their reproduction is by having male kids who are alpha males but in these circumstances an alpha is someone who is good at seduction rather than the old style coercion and multiple wives ownership of old times.

tl;dr the bottleneck for overpopulation is individual women's fertility, and the way women maximize their reproduction is by having high quality sons rather than popping out babies nonstop. So you can still have high reproductive strategies without actual overpopulation.

In any case it's hard to think in these terms, the feeling I have is memetics will always overshadow any purely instinctual drives.

Comment by Bugle on Non-Malthusian Scenarios · 2009-09-28T13:39:40.753Z · LW · GW

But the new universes also have their own population, though I guess you could colonize universes where humans don't arise rather than universes identical to this one except I didn't scratch my nose just now

Comment by Bugle on Non-Malthusian Scenarios · 2009-09-26T23:01:45.744Z · LW · GW

How about simple spontaneous population stability... I live in a country with negative birth rate but the population is increasing due to immigration nevertheless. This state of affairs hasn't been legislated into existence, it just happened, and may be a natural behaviour of large human populations. Perhaps once the whole world reaches western standards of living the whole world will stop growing exponentially, with pockets of negative growth being compensated by low but positive growth in others... in the long term the trend could even be for decreasing population...

Comment by Bugle on The Finale of the Ultimate Meta Mega Crossover · 2009-09-20T19:23:59.130Z · LW · GW

I guess I'll be back once I've read Permutation city then...

Comment by Bugle on Torture vs. Dust Specks · 2009-09-12T18:00:53.787Z · LW · GW

Although if we factor in consequences, say... being distracted by a dust speck in the eye while driving or doing any other such critical activity then statistically those trillions of dust specks have the potential to cause untold amounts of damage and suffering

Comment by Bugle on Torture vs. Dust Specks · 2009-09-12T13:05:05.835Z · LW · GW

There is a false choice being offered, because every person in every lifetime is going to experience getting something in their eye, I get a bug flying into my eye on a regular basis whenever I go running (3 of them the last time!) and it'll probably have happened thousands of times to me at the end of my life. It's pretty much a certainty of human experience (Although I suppose it's statistically possible for some people to go through life without ever getting anything in their eyes).

Is the choice being offered to make all humanities eyes for all eternity immune to small inconveniences such as bugs, dust or eyelashes? Otherwise we really aren't being offered anything at all.

Comment by Bugle on Why You're Stuck in a Narrative · 2009-08-06T02:34:25.648Z · LW · GW

But remember like Alan Moore said "The one place where gods exist unquestionably is the human mind". Similarly, narratives are a fact of not just your brain, but that of everyone around you, realizing this can be convenient.