Posts

Comments

Comment by davekasten on Nathan Helm-Burger's Shortform · 2024-04-19T03:44:01.537Z · LW · GW

Like, I hear you, but that is...also not how they teach gun safety.  Like, if there is one fact you know about gun safety, it's that the entire field emphasizes that a gun is inherently dangerous towards anything it is pointed towards.

Comment by davekasten on Akash's Shortform · 2024-04-18T16:42:56.807Z · LW · GW

I largely agree, but think given government hiring timelines, there's no dishonor in staying at a lab doing moderately risk-reducing work until you get a hiring offer with an actual start date.  This problem is less bad for the special hiring authorities being used for AI stuff oftentimes, but it's still not ideal.

Comment by davekasten on Mid-conditional love · 2024-04-17T19:35:37.834Z · LW · GW

Oh wow, this is...not what I thought people meant when they say "unconditional love." 

In my circles, "conditional love" is about love with lots of threats and demands that the other person change, and if failing to do so, they would be told they're unworthy of being loved by their partner.

Comment by davekasten on What convincing warning shot could help prevent extinction from AI? · 2024-04-16T20:32:35.144Z · LW · GW

As you know from our conversations, I'm largely in the same camp as you on this point.  

But one point I'd make incrementally is this: USG folks are also concerned about warning shots of the nature, "The President's Daily Brief ran an article 6 months ago saying warning signs for dangerous thing X would be events W, Y, and Z, and today the PDB had an article saying our intelligence agencies assess that Y and Z have happened due to super secret stuff".

If rationalists want rationalist warning shots to be included, they need to convince relevant government analytic stakeholders of their relevance. 

Comment by davekasten on nikola's Shortform · 2024-04-15T15:24:38.470Z · LW · GW

That's probably true if the takeover is to maximize the AI's persistence.  You could imagine a misaligned AI that doesn't care about its own persistence -- e.g., an AI that got handed a misformed min() or max() that causes it to kill all humans instrumental to its goal (e.g., min(future_human_global_warming))

Comment by davekasten on LessWrong's (first) album: I Have Been A Good Bing · 2024-04-01T16:24:26.019Z · LW · GW

The only rational response to AI music generation is live-only music creation within the context of traditional norms that prohibit the use of electricity for productive work.

Unrelatedly, I am co-organizing a kabbalat shabbat at Manifest Conference this year

Comment by davekasten on Scale Was All We Needed, At First · 2024-03-25T19:29:14.568Z · LW · GW

The fridge horror of this story comes when you realize that the AI timed the call to Biden until it was sure that Biden's advisor had been successfully and fully briefed

Comment by davekasten on The World in 2029 · 2024-03-11T16:15:15.785Z · LW · GW

Market on 2029 culture war: 

Comment by davekasten on Are we so good to simulate? · 2024-03-05T16:54:39.621Z · LW · GW

I think I'm confused about something in your reference-class argument here.  I'm going to say it in a very boring way, on the assumption someone will come by and tell me how I'm wrong shortly:   It seems like you're saying that the right reference class is "all possible minds". But isn't the right reference class "minds that look like humans in the 21st century at the (putative) dawn of AI"?  

Separately, do we know that we expect all sims to have equal distributions of real to unreal persons as we believe to be the case?  I don't think I've ever met someone from lots of countries; why couldn't I be in the ancestor sim where most persons are p-zombies running on simplified scripts unless they're needed to suddenly have detailed interiority "for the plot"?

Comment by davekasten on The World in 2029 · 2024-03-04T19:24:43.192Z · LW · GW

I find the subscripts distracting, personally, and would prefer something like "Zebras discover fire (60%)" in parentheses

Comment by davekasten on The World in 2029 · 2024-03-04T04:59:04.258Z · LW · GW

Ok, I'll bite: does that mean that you think China is less likely to invade China if recession?  This would be happy news, but is not necessarily consensus view among DC defense wonks (i.e. the fear of someone saying, "We need a Short Victorious War").

Comment by davekasten on Retirement Accounts and Short Timelines · 2024-02-22T15:04:51.264Z · LW · GW

I mean, does your Vanguard targeted lifecycle index fund likely invest in equities exposed to AGI growth (conditional on non-doom)?  

If you think money still has meaning after AGI and meaningful chance of no-doom, it might actually be optimal to invest in your retirement fund.

Comment by davekasten on Things You’re Allowed to Do: University Edition · 2024-02-19T04:36:42.796Z · LW · GW

I feel like, "adopt a religious/moral stringency or observance" or stopping the same is also a key item here.

Comment by davekasten on Guide to rationalist interior decorating · 2024-02-15T14:49:11.987Z · LW · GW

I would like to thank you for this post and for encouraging me to get a ball of fairy lights, which provides me with a genuinely unexpected amount of joy whenever I look at it up close.