Preparing for the Intelligence Explosion

post by fin, wdmacaskill · 2025-03-11T15:38:29.524Z · LW · GW · 8 comments

This is a link post for https://www.forethought.org/research/preparing-for-the-intelligence-explosion

Contents

8 comments

This is a linkpost for a new paper called Preparing for the Intelligence Explosion, by Will MacAskill and Fin Moorhouse. It sets the high-level agenda for the sort of work that Forethought is likely to focus on.

Some of the areas in the paper that we expect to be of most interest to EA Forum or LessWrong readers are:

Here’s the abstract:

AI that can accelerate research could drive a century of technological progress over just a few years. During such a period, new technological or political developments will raise consequential and hard-to-reverse decisions, in rapid succession. We call these developments grand challenges. 

These challenges include new weapons of mass destruction, AI-enabled autocracies, races to grab offworld resources, and digital beings worthy of moral consideration, as well as opportunities to dramatically improve quality of life and collective decision-making.

We argue that these challenges cannot always be delegated to future AI systems, and suggest things we can do today to meaningfully improve our prospects. AGI preparedness is therefore not just about ensuring that advanced AI systems are aligned: we should be preparing, now, for the disorienting range of developments an intelligence explosion would bring.

8 comments

Comments sorted by top scores.

comment by habryka (habryka4) · 2025-03-11T20:03:31.346Z · LW(p) · GW(p)

My guess is you know this, but the sidenote implementation appears to be broken. When clicking on the footnote labeled "1" it opens up a footnote labeled "2", and also, the footnotes overlap on the right in very broken looking ways: 

Replies from: wdmacaskill, max-dalton
comment by wdmacaskill · 2025-03-12T08:49:21.909Z · LW(p) · GW(p)

Thanks - appreciate that! It comes up a little differently for me, but still an issue - we've asked the devs to fix. 

comment by Max Dalton (max-dalton) · 2025-03-12T08:40:44.034Z · LW(p) · GW(p)

Thanks Oli! I think the clustering issue is fixed now, looking into what's going on with the numbers.

comment by Julian Bradshaw · 2025-03-12T02:59:29.825Z · LW(p) · GW(p)

Meta: I'm kind of weirded out by how apparently everyone is making their own high-effort custom-website-whitepapers? Is this something that's just easier with LLMs now? Did Situational Awareness create a trend? I can't read all this stuff man.

In general there seems to be way more high-effort work coming out since reasoning models got released. Maybe it's just crunchtime.

Replies from: wdmacaskill
comment by wdmacaskill · 2025-03-12T08:53:37.073Z · LW(p) · GW(p)

There's definitely a new trend towards custom-website essays. Forethought is a website for lots of research content, though (like Epoch), not just PrepIE.

And I don't think it's because of people getting more productive because of reasoning models - AI was helpful for PrepIE but more like 10-20% productivity boost than 100% boost, and I don't think AI was used much for SA, either.

comment by Immanuel Jankvist (emanueljankvist) · 2025-03-11T22:46:40.791Z · LW(p) · GW(p)

The following seems a bit unclear to me, and might warrant an update–if I am not alone in the assessment:

Section 3 finds that even without a software feedback loop (i.e. “recursive self-improvement”), [...], then we should still expect very rapid technological development [...] once AI meaningfully substitutes for human researchers.

I might just be taking issue with the word "without" and taking it in a very literal sense, but to me "AI meaningfully substituting for human researchers" implies at least a weak form of recursive self-improvement.
That is, I would be quite surprised if the world allowed for AI to become as smart as human researchers but no smarter afterwards.

Replies from: wdmacaskill, caleb-biddulph
comment by wdmacaskill · 2025-03-12T09:00:42.455Z · LW(p) · GW(p)

Ah, by the "software feedback loop" I mean: "At the point of time at which AI has automated AI R&D, does a doubling of cognitive effort result in more than a doubling of output? If yes, there's a software feedback loop - you get (for a time, at least) accelerating rates of algorithmic efficiency progress, rather than just a one-off gain from automation." 

I see now why you could understand "RSI" to mean "AI improves itself at all over time". But even so, the claim would still hold - even if (implausibly) AI gets no smarter than human-level, you'd still get accelerated tech development,  because the quantity of AI research effort would increase at a growth rate much faster than the quantity of human research effort. 

comment by CBiddulph (caleb-biddulph) · 2025-03-12T00:38:58.493Z · LW(p) · GW(p)

I interpreted this as "even without a software feedback loop, there will be very rapid technological development; this gives a lower bound on the actual pace of technological development, since there will almost certainly be some feedback loop"