post by [deleted] · · ? · GW · 0 comments

This is a link post for

0 comments

Comments sorted by top scores.

comment by Nathan Helm-Burger (nathan-helm-burger) · 2023-05-17T19:16:53.957Z · LW(p) · GW(p)
  • Do prosaic “bandaid” solutions like this appeal to you? (why/why not?)
  • Relate your answer to the previous question to your perspective on takeoff probabilities (if possible, I would be very interested in hearing actual numbers for the likelihood of certain takeoff scenarios).

 

I think we need both. I think if we have only bandaid solutions, then we've only bought ourselves a few years delay in doom. I think if we don't have bandaid solutions, we won't buy ourselves enough time for the more robust solutions to have a chance of coming up with workable practical solutions.

Taking as a premise that sometime in the next 4 years some AI lab / research group makes sufficient advances to begin using near-current-level AI to substantially speed up their development of future generations of AI. Timeframes given are measured from start of this process of heavily AI-supported model development. I'm about 90% confident that this premise will be the case.

I think a hard take-off in the course of < 3 months which goes from near-current-level AI to overwhelmingly superhumanly powerful AGI is highly implausible. I'd personally assign a less than 1% chance to that.

I think a medium-soft take-off of near-current-level AI to 100x more powerful AGI over the course of > 3 months but < 18 months is highly plausible and fairly likely unless substantial regulatory efforts are made to prevent this. I'd give something like 60% chance of this happening whether or not an attempt at regulation is made.

I think soft take-off of near-current-level AI to 100x more powerful AGI over the course of > 18 months but < 5 years as taking up most of the rest of my probability in this space. I'd say that I think this is likely to be the shape of the improvement curve only in the face of successful regulation to slow down the process. I'd give that around 30% probability.

The remaining probability is in 'I dunno, something really weird happens, predicting the future is hard.'

comment by Nathan Helm-Burger (nathan-helm-burger) · 2023-05-17T19:01:33.141Z · LW(p) · GW(p)

inevitability of

 

I suggest a change to more epistemologically robust phasing 'my high level of credence in'

Replies from: marc/er
comment by lukemarks (marc/er) · 2023-05-18T08:35:23.039Z · LW(p) · GW(p)

Agreed and edited.