Posts

“AI Risk Discussions” website: Exploring interviews from 97 AI Researchers 2023-02-02T01:00:01.067Z

Comments

Comment by zchuang on What Does LessWrong/EA Think of Human Intelligence Augmentation as of mid-2023? · 2023-07-08T23:20:42.109Z · LW · GW

Sorry but aren't we in a fast takeoff world at the point of WBE.  What's the disjunctive world of no recursive self-improvement and WBE? 

Comment by zchuang on Statement on AI Extinction - Signed by AGI Labs, Top Academics, and Many Other Notable Figures · 2023-06-06T06:23:54.317Z · LW · GW

He posted on a twitter a request to talk to people who feel strongly here.

Comment by zchuang on Clarifying and predicting AGI · 2023-05-10T16:46:14.171Z · LW · GW

Yeah, re-reading I realise I was unclear. Given your claim: "by the time we get to 2000 in that, such AGIs will be automating huge portions of AI R&D,". I'm asking the following:

  1. Is the 2000 mark predicated on automation of things we can't envision now (finding secret sauce to singularity) or is it predicated off pushing existing things like AI R&D finds better compute or is it a combination of both?
  2. What's the empirical on the ground representative modal action you're seeing at 2025 from either your vignette (e.g. I found the diplomacy AI super important for grokking what short timelines were to me). I guess it's more asking what you see as the divergence between you and Richard at 2025 that's represented by the difference of 25 and 100.

Hopefully that made the questions clearer.

Comment by zchuang on Clarifying and predicting AGI · 2023-05-10T14:59:25.685Z · LW · GW

Sorry for a slightly dumb question but in your part of the table you set 2000 as the year before singularity and your explanation is that 2000-second tasks jump to singularity. Is your model of fast take-off then contingent on there being more special sauce for intelligence being somewhat redundant as a crux because recursive self-improvement is just much more effective. I'm having trouble envisioning a 2000-second task + more scaling and tuning --> singularity. 

Additional question is what your model of falsification is for let's say 25-second task vs. 100-second task in 2025 because it seems like reading your old vignettes you really nailed the diplomacy AI part.

Also slightly pedantic but there's a typo on 2029 on Richard's guess.