How likely is AGI to force us all to be happy forever? (much like in the Three Worlds Collide novel)

post by uhbif19 · 2025-01-18T15:39:21.549Z · LW · GW · No comments

This is a question post.

Contents

  Answers
    2 Dagon
    2 jbash
None
No comments

Hi, everyone. I'm not sure if my post is well-written, but I think LW might be the only right place to have this discussion. Feel free to suggest changes.


AGI may arrive soon and it is possible that it would kill us all. This does not bother me that much, as dying happens to people all the time.
But losing control of our mind and lifespan is a much less pleasant perspective.

As far as I am aware, people usually do not let others kill themselves. In some situations, this may lead to involuntary placement in a mental hospital. And as not living forever may be seen as a form of killing yourself, AGI may quite well not let you have a finite lifespan. That places you in the uncomfortable situation of being trapped with AGI forever.

Going further. We do not know what peak happiness looks like, and our regular state may be very different from it. And as EY outlined in Three Worlds Collide [LW · GW], letting us live our miserable lives may be unacceptable. This change may not align well with our understanding of life's purpose. And being trapped to live that way forever might not be that desirable.


So my question is: would you see that as a possible outcome? If so, do you see a way to somehow see that happening in advance?

Answers

answer by Dagon · 2025-01-18T20:25:05.532Z · LW(p) · GW(p)

You need to be clear who is included in "us".  AI is likely to be trained on human understanding of identity and death, which is very much based on generational replacement rather than continuity over centuries.  Some humans wish this wasn't so, and hope it won't apply to them, but there's not enough examples (none in truth, few and unrealistic in fiction) to train on or learn from.

It seems likely that if "happy people" ends up in the AI goalset, it'll create new ones that have higher likelihood of being happy than those in the past.  Honestly, I'm going to be dead, so my preference doesn't carry much weight, but I think I prefer to imagine tiling the universe with orgasmium more than I do paperclips.

It's FAR more effort to make an existing damaged human (as all are in 2025) happy than just to make a new happy human.

answer by jbash · 2025-01-18T18:37:00.016Z · LW(p) · GW(p)

would you see that as a possible outcome?

Sure. Not the most likely outcome, but not so improbable as all that.

Reservation: ASI (what you're suggesting is beyond mere AGI) will still exist in the physical world and have physical limitations, so you will eventually die anyway. But it could be a very long time.

If so, do you see a way to somehow see that happening in advance?

Not really, no.

Not beyond obvious stuff like watching what's being built and how, and listening to what it and its creators say about its values.

And as not living forever may be seen as a form of killing yourself, AGI may quite well not let you have a finite lifespan. That places you in the uncomfortable situation of being trapped with AGI forever.

Yes, that's one of many kinds of lock-in of present human values that could go wrong. But, hey, it'd be aligned, I guess.

The no-killing-yourself rule isn't completely universal, though, and there's dissent around it, and it's often softened to "no killing yourself unless I agree your situation sucks enough", or even the more permissive "no killing yourself unless your desire to do so is clearly a fixed, stable thing".

I actually think there's a less than 50-50 chance that people would intentionally lock in the hardest form of the rule permanently if they were consciously defining the AI's goals or values.

We do not know what peak happiness looks like, and our regular state may be very different from it. And as EY outlined in Three Worlds Collide, letting us live our miserable lives may be unacceptable.

That seems like a very different question from the one about indefinite life extension. Life extension isn't the main change the Supperhappies make in that story[1].

This change may not align well with our understanding of life's purpose. And being trapped to live that way forever might not be that desirable.

Pre-change understanding, or post-change understanding? Desirable to pre-change you, or to post-change you?

If you see your pre-change values as critical parts of Who You Are(TM), and they get rewritten, then aren't you effectively dead anyway, with your place taken by a different person with different values, who's actually pretty OK with the whole thing? If being dead doesn't worry you, why would that worry you?


  1. In fact, the humans had very long lives going into the story, and I don't remember anything in the story that actually said humans weren't enforcing a no-killing-yourself rule among themselves up to the very end. ↩︎

comment by Walker Vargas · 2025-01-19T00:19:24.375Z · LW(p) · GW(p)

In the ending where humanity gets modified, some people commit suicide. The captain thinks it doesn't make sense to choose complete erasure over modification.

No comments

Comments sorted by top scores.