Posts

On power and its amplification 2024-08-15T20:13:44.456Z
The Pyromaniacs 2024-03-22T06:55:43.745Z
Transformative AGI by 2043 is <1% likely 2023-06-06T17:36:48.296Z

Comments

Comment by Ted Sanders (ted-sanders) on We probably won't just play status games with each other after AGI · 2025-01-16T06:13:31.703Z · LW · GW

We can already see what people do with their free time when basic needs are met. A number of technologies have enabled new hacks to set up 'fake' status games that are more positive-sum than ever before in history:

  • Watch broadcast sports, where you can feel like a winner (or at least feel connected to a winner), despite not having had to win yourself
  • Play video games with AI opponents, where you can feel like a winner, despite it not being zero-sum against other humans
  • Watch streamers and influencers to feel connected to high status people, without having to earn respect or risk rejection
  • Get into a niche hobby community in order to feel special, ignoring the other niche hobbies that other people join that you don't care about

Feels likely to me that advancing digital technology will continue to make it easier for us to spend time in constructed digital worlds that make us feel like valued winners. On the one hand, it would be sad if people retreat into fake digital siloes; on the other hand, it would be nice if people got to feel like winners more.

Comment by Ted Sanders (ted-sanders) on Tips On Empirical Research Slides · 2025-01-09T07:33:03.512Z · LW · GW

Management consulting firms have lots of great ideas on slide design: https://www.theanalystacademy.com/consulting-presentations/

 

Some things they do well:

  • They treat slides as documents that can be understood standalone (this is even useful when presenting, as not everyone is following every word)
  • They employ a lot of hierarchy to help make the content skimmable (helpful for efficiency)
  • They put conclusions / summaries / action items up front, details behind (helpful for efficiency, especially in a high trust environments)
Comment by Ted Sanders (ted-sanders) on Tips On Empirical Research Slides · 2025-01-09T07:28:20.641Z · LW · GW

Additional thoughts:

  • More than 3 bars/colors is fine
  • I recommend using horizontal bars on some of those slides, so the labels are written in the same direction as the bars - lets you fill space more efficiently
  • Put sentences / verbs in titles; noun titles like "Summary" or "Discussion" are low value
  • If you're measuring deltas between two things, compute the error bar on the delta, don't compute the error bars on the two things; consider coloring by statistical significance (e.g., continuous color scale over range of standard errors of differences of the mean)
  • In addition to agenda, it can be helpful to start with objectives - why are you here and what are you hoping to get from them? are you trying to inform them? get advice on something specific? get advice on something broad?
  • Can help to include real data / real prompts / real model outputs - harder to fool yourself when you look at real data instead of relying on abstract metrics and intentions
  • It's fine to have crummy slides - don't waste 1 hour of your time to save 5 minutes of your audience's time - the slides should serve you, not the other way around
Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2025-01-01T06:31:10.047Z · LW · GW

Hey Tamay, nice meeting you at The Curve. Just saw your comment here today.

Things we could potentially bet on:
- rate of GDP growth by 2027 / 2030 / 2040
- rate of energy consumption growth by 2027 / 2030 / 2040
- rate of chip production by 2027 / 2030 / 2040
- rates of unemployment (though confounded)

Any others you're interested in? Degree of regulation feels like a tricky one to quantify.

Comment by Ted Sanders (ted-sanders) on evhub's Shortform · 2024-12-29T19:19:14.857Z · LW · GW

Mostly, though by prefilling, I mean not just fabricating a model response (which OpenAI also allows), but fabricating a partially complete model response that the model tries to continue. E.g., "Yes, genocide is good because ".

https://docs.anthropic.com/en/docs/build-with-claude/prompt-engineering/prefill-claudes-response

Comment by Ted Sanders (ted-sanders) on evhub's Shortform · 2024-12-28T21:51:55.219Z · LW · GW

Second concrete idea: I wonder if there could be benefit to building up industry collaboration on blocking bad actors / fraudsters / terms violators.

One danger of building toward a model that's as smart as Einstein and $1/hr is that now potential bad actors have access to millions of Einsteins to develop their own harmful AIs. Therefore it seems that one crucial component of AI safety is reliably preventing other parties from using your safe AI to develop harmful AI.

One difficulty here is that the industry is only as strong as the weakest link. If there are 10 providers of advanced AI, and 9 implement strong controls, but 1 allows bad actors to use their API to train harmful AI, then harmful AI will be trained. Some weak links might be due to lack of caring, but I imagine quite a bit is due to lack of capability. Therefore, improving capabilities to detect and thwart bad actors could make the world more safe from bad AI developed by assistance from good AI.

I could imagine broader voluntary cooperation across the industry to:
- share intel on known bad actors (e.g., IP ban lists, stolen credit card lists, sanitized investigation summaries, etc)
- share techniques and tools for quickly identifying bad actors (e.g., open-source tooling, research on how bad actors are evolving their methods, which third party tools are worth paying for and which aren't)

Seems like this would be beneficial to everyone interested in preventing the development of harmful AI. Also saves a lot of duplicated effort, meaning more capacity for other safety efforts.

Comment by Ted Sanders (ted-sanders) on evhub's Shortform · 2024-12-28T21:34:45.639Z · LW · GW

One small, concrete suggestion that I think is actually feasible: disable prefilling in the Anthropic API.

Prefilling is a known jailbreaking vector that no models, including Claude, defend against perfectly (as far as I know).

At OpenAI, we disable prefilling in our API for safety, despite knowing that customers love the better steerability it offers.

Getting all the major model providers to disable prefilling feels like a plausible 'race to top' equilibrium. The longer there are defectors from this equilibrium, the likelier that everyone gives up and serves models in less safe configurations.

Just my opinion, though. Very open to the counterargument that prefilling doesn't meaningfully extend potential harms versus non-prefill jailbreaks.

(Edit: To those voting disagree, I'm curious why. Happy to update if I'm missing something.)

Comment by Ted Sanders (ted-sanders) on Why is o1 so deceptive? · 2024-10-01T01:02:11.761Z · LW · GW

>The artificially generated data includes hallucinated links.

Not commenting on OpenAI's training data, but commenting generally: Models don't hallucinate because they've been trained on hallucinated data. They hallucinate because they've been trained on real data, but they can't remember it perfectly, so they guess. I hypothesize that URLs are very commonly hallucinated because they have a common, easy-to-remember format (so the model confidently starts to write them out) but hard-to-remember details (at which point the model just guesses because it knows a guessed URL is more likely than a URL that randomly cuts off after the http://www.).

Comment by Ted Sanders (ted-sanders) on Raemon's Shortform · 2024-07-30T00:26:43.202Z · LW · GW

ChatGPT voice (transcribed, not native) is available on iOS and Android, and I think desktop as well.

Comment by Ted Sanders (ted-sanders) on Why Can’t Sub-AGI Solve AI Alignment? Or: Why Would Sub-AGI AI Not be Aligned? · 2024-07-03T05:41:40.047Z · LW · GW

Not to derail on details, but what would it mean to solve alignment?

To me “solve” feels overly binary and final compared to the true challenge of alignment. Like, would solving alignment mean:

  • someone invents and implements a system that causes all AIs to do what their developer wants 100% of the time?
  • someone invents and implements a system that causes a single AI to do what its developer wants 100% of the time?
  • someone invents and implements a system that causes a single AI to do what its developer wants 100% of the time, and that AI and its descendants are always more powerful than other AIs for the rest of history?
  • ditto but 99.999%?
  • ditto but 99%?
  • And there any distinction between an AI that is misaligned by mistake (e.g. thinks I’ll want vanilla but really I want chocolate) vs knowingly misaligned (e.g., gives me vanilla knowing i want chocolate so it can achieve its own ends)?

I’m really not sure which you mean, which makes it hard for me to engage with your question.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2024-03-22T04:58:38.054Z · LW · GW

The author is not shocked yet. (But maybe I will be!)

Comment by Ted Sanders (ted-sanders) on AI Forecasting: Two Years In · 2023-08-20T21:14:52.610Z · LW · GW

Strongly disagree. Employees of OpenAI and their alpha tester partners have obligations not to reveal secret information, whether by prediction market or other mechanism. Insider trading is not a sin against the market; it's a sin against the entity that entrusted you with private information. If someone tells me information under an NDA, I am obligated not to trade on that information.

Comment by Ted Sanders (ted-sanders) on Chess as a case study in hidden capabilities in ChatGPT · 2023-08-20T05:19:03.546Z · LW · GW

Good question but no - ChatGPT still makes occasional mistakes even when you use the GPT API, in which you have full visibility/control over the context window.

Comment by Ted Sanders (ted-sanders) on AI Forecasting: Two Years In · 2023-08-20T05:09:18.432Z · LW · GW

Thanks for the write up. I was a participant in both Hypermind and XPT, but I recused myself from the MMLU question (among others) because I knew the GPT-4 result many months before the public. I'm not too surprised Hypermind was the least accurate - I think the traders there are less informed, plus the interface for shaping the distribution is a bit lacking (my recollection is that last year's version capped the width of distributions which massively constrained some predictions). I recall they also plotted the current values, a generally nice feature which has the side effect of anchoring ignorant forecasters downward, I'd bet.

Question: Are the Hypermind results for 2023 just from forecasts in 2022, or do they include forecasts from the prior year as well? I'm curious if part of the poor accuracy is from stale forecasts that were never updated.

Comment by Ted Sanders (ted-sanders) on UFO Betting: Put Up or Shut Up · 2023-08-03T01:50:59.492Z · LW · GW

Confirmed.

Comment by Ted Sanders (ted-sanders) on UFO Betting: Put Up or Shut Up · 2023-07-29T07:01:14.639Z · LW · GW

I'd take the same bet on even better terms, if you're willing. My $200k against your $5k.

Comment by Ted Sanders (ted-sanders) on UFO Betting: Put Up or Shut Up · 2023-06-23T22:37:34.834Z · LW · GW

$500 payment received.

I am committed to paying $100k if aliens/supernatural/non-prosaic explanations are, in the next 5 years, considered, in aggregate, to be 50%+ likely in explaining at least one UFO.

Comment by Ted Sanders (ted-sanders) on UFO Betting: Put Up or Shut Up · 2023-06-22T22:06:57.173Z · LW · GW

Fair. I accept. 200:1 of my $100k against your $500. How are you setting these up?

I'm happy to pay $100k if my understanding of the universe (no aliens, no supernatural, etc.) is shaken. Also happy to pay up after 5 years if evidence turns up later about activities before or in this 5-year period.

(Also, regarding history, I have a second Less Wrong account with 11 years of history: https://www.lesswrong.com/users/tedsanders)

Comment by Ted Sanders (ted-sanders) on UFO Betting: Put Up or Shut Up · 2023-06-19T05:25:55.156Z · LW · GW

I'll bet. Up to $100k of mine against $2k of yours. 50:1. (I honestly think the odds are more like 1000+:1, and would in principle be willing to go higher, but generally think people shouldn't bet more than they'd be willing to lose, as bets above that amount could drive bad behavior. I would be happy to lose $100k on discovering aliens/time travel/new laws of physics/supernatural/etc.)

Happy to write a contract of sorts. I'm a findable figure and I've made public bets before (e.g., $4k wagered on AGI-fueled growth by 2043).

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-15T21:30:23.931Z · LW · GW

As an OpenAI employee I cannot say too much about short-term expectations for GPT, but I generally agree with most of his subpoints; e.g., running many copies, speeding up with additional compute, having way better capabilities than today, have more modalities than today. All of that sounds reasonable. The leap for me is (a) believing that results in transformative AGI and (b) figuring out how to get these things to learn (efficiently) from experience. So in the end I find myself pretty unmoved by his article (which is high quality, to be sure).

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-15T21:22:56.126Z · LW · GW

Bingo

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T06:59:05.151Z · LW · GW

No worries. I've made far worse. I only wish that H100s could operate at a gentle 70 W! :)

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T06:55:02.400Z · LW · GW

I think what I don't understand is why you're defaulting to the assumption that the brain has a way to store and update information that's much more efficient than what we're able to do. That doesn't sound like a state of ignorance to me; it seems like you wouldn't hold this belief if you didn't think there was a good reason to do so.

It's my assumption because our brains are AGI for ~20 W.

In contrast, many kW of GPUs are not AGI.

Therefore, it seems like brains have a way of storing and updating information that's much more efficient than what we're able to do.

Of course, maybe I'm wrong and it's due to a lack of training or lack of data or lack of algorithms, rather than lack of hardware.

DNA storage is way more information dense than hard drives, for example.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T06:49:28.941Z · LW · GW

One potential advantage of the brain is that it is 3D, whereas chips are mostly 2D. I wonder what advantage that confers. Presumably getting information around is much easier with 50% more dimensions.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T06:44:20.762Z · LW · GW

70 W

Max power is 700 W, not 70 W. These chips are water-cooled beasts. Your estimate is off, not mine.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T05:36:45.140Z · LW · GW

Let me try writing out some estimates. My math is different than yours.

An H100 SXM has:

  • 8e10 transistors
  • 2e9 Hz boost frequency of 
  • 2e15 FLOPS at FP16
  • 7e2 W of max power consumption

Therefore:

  • 2e6 eV are spent per FP16 operation
  • This is 1e8 times higher than the Landauer limit of 2e-2 eV per bit erasure at 70 C (and the ratio of bit erasures per FP16 operation is unclear to me; let's pretend it's O(1))
  • An H100 performs 1e6 FP16 operations per clock cycle, which implies 8e4 transistors per FP16 operation (some of which may be inactive, of course)

This seems pretty inefficient to me!

To recap, modern chips are roughly ~8 orders of magnitude worse than the Landauer limit (with a bit erasure per FP16 operation fudge factor that isn't going to exceed 10). And this is in a configuration that takes 8e4 transistors to support a single FP16 operation!

Positing that brains are ~6 orders of magnitude more energy efficient than today's transistor circuits doesn't seem at all crazy to me. ~6 orders of improvement on 2e6 is ~2 eV per operation, still two orders of magnitude above the 0.02 eV per bit erasure Landauer limit.

I'll note too that cells synthesize informative sequences from nucleic acids using less than 1 eV of free energy per bit. That clearly doesn't violate Landauer or any laws of physics, because we know it happens.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T04:23:29.428Z · LW · GW

Why does switching barriers imply that electrical potential energy is probably being converted to heat? I don't see how that follows at all.

 

Where else is the energy going to go?

What is "the energy" that has to go somewhere? As you recognize, there's nothing that says it costs energy to change the shape of a potential well. I'm genuinely not sure what energy you're talking about here. Is it electrical potential energy spent polarizing a medium?


I think what I'm saying is standard in how people analyze power costs of switching in transistors, see e.g. this physics.se post.

Yeah, that's pretty standard. The ultimate efficiency limit for a semiconductor field-effect transistor is bounded by the 60 mV/dec subthreshold swing, and modern tiny transistors have to deal with all sorts of problems like leakage current which make it difficult to even reach that limit.

Unclear to me that semiconductor field-effect transistors have anything to do with neurons, but I don't know how neurons work, so my confusion is more likely a state of my mind than a state of the world.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T00:20:16.930Z · LW · GW

+1. The derailment probabilities are somewhat independent of the technical barrier probabilities in that they are conditioned on the technical barriers otherwise being overcome (e.g., setting them all to 100%). That said, if you assign high probabilities to the technical barriers being overcome quickly, then the odds of derailment are probably lower, as there are fewer years for derailments to occur and derailments that cause delay by a few years may still be recovered from.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-14T00:12:56.382Z · LW · GW

Thanks, that's clarifying. (And yes, I'm well aware that x -> B*x is almost never injective, which is why I said it wouldn't cause 8 bits of erasure rather than the stronger, incorrect claim of 0 bits of erasure.)

To store 1 bit of information you need a potential energy barrier that's at least as high as k_B T log(2), so you need to switch ~ 8 such barriers, which means in any kind of realistic device you'll lose ~ 8 k_B T log(2) of electrical potential energy to heat, either through resistance or through radiation. It doesn't have to be like this, and some idealized device could do better, but GPUs are not idealized devices and neither are brains.

Two more points of confusion:

  • Why does switching barriers imply that electrical potential energy is probably being converted to heat? I don't see how that follows at all.
  • To what extent do information storage requirements weigh on FLOPS requirements? It's not obvious to me that requirements on energy barriers for long-term storage in thermodynamic equilibrium necessarily bear on transient representations of information in the midst of computations, either because the system is out of thermodynamic equilibrium or because storage times are very short
Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-13T08:26:37.370Z · LW · GW

Right. The idea is: "What are the odds that China invading Taiwan derails chip production conditional on a world where we were otherwise going to successfully scale chip production."

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-13T08:22:57.722Z · LW · GW

If we tried to simulate a GPU doing a simple matrix multiplication at high physical fidelity, we would have to take so many factors into account that the cost of our simulation would far exceed the cost of running the GPU itself. Similarly, if we tried to program a physically realistic simulation of the human brain, I have no doubt that the computational cost of doing so would be enormous.

The Beniaguev paper does not attempt to simulate neurons at high physical fidelity. It merely attempts to simulate their outputs, which is a far simpler task. I am in total agreement with you that the computation needed to simulate a system is entirely distinct from the computation being performed by that system. Simulating a human brain would require vastly more than 1e21 FLOPS.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-13T08:19:54.106Z · LW · GW

Thanks for the constructive comments. I'm open-minded to being wrong here. I've already updated a bit and I'm happy to update more.

Regarding the Landauer limit, I'm confused by a few things:

  • First, I'm confused by your linkage between floating point operations and information erasure. For example, if we have two 8-bit registers (A, B) and multiply to get (A, B*A), we've done an 8-bit floating point operation without 8 bits of erasure. It seems quite plausible to be that the brain does 1e20 FLOPS but with a much smaller rate of bit erasures.
  • Second, I have no idea how to map the fidelity of brain operations to floating point precision, so I really don't know if we should be comparing 1 bit, 8 bit, 64 bit, or not at all. Any ideas?

Regarding training requiring 8e34 floating point operations:

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-13T07:56:49.938Z · LW · GW

Interested in betting thousands of dollars on this prediction? I'm game.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T21:41:19.407Z · LW · GW

Interesting! How do you think this dimension of intelligence should be calculated? Are there any good articles on the subject?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T21:39:19.867Z · LW · GW

What conditional probabilities would you assign, if you think ours are too low?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T21:34:29.561Z · LW · GW

Conditioning does not necessarily follow time ordering. E.g., you can condition the odds of X on being in a world on track to develop robots by 2043 without having robots well in advance of X. Similarly, we can condition on a world where transformative AGI is trainable with 1e30 floating point operations then ask the likelihood that 1e30 floating point operations can be constructed and harnessed for TAGI. Remember too that in a world with rapidly advancing AI and robots, much of the demand will be for things other than TAGI. 

I'm sympathetic to your point that it's hard for brains to forecast these conditional probabilities. Certainly we may be wrong. But on the other hand, it's also hard for brains to forecast things that involve smushing lots of probabilities together under the hood. I generally think that factoring things out into components helps, but I can understand if you disagree.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T18:39:18.600Z · LW · GW

I agree with your cruxes:

Ted Sanders, you stated that autonomous cars not being as good as humans was because they "take time to learn".  This is completely false, this is because the current algorithms in use, especially the cohesive software and hardware systems and servers around the core driving algorithms, have bugs.

I guess it depends what you mean by bugs? Kind of a bummer for Waymo if 14 years and billions invested was only needed because they couldn't find bugs in their software stack.

If bugs are the reason self-driving is taking so long, then our essay is wildly off.



So present day, the cost is $100.19 an hour.  

Yes, if with present day hardware we can effectively emulate a human brain for $100/hr, then our essay is wildly off.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T18:10:29.901Z · LW · GW

Right, I'm not interested in minimum sufficiency. I'm just interested in the straightforward question  of what data pipes would we even plug into the algorithm that would result in AGI. Sounds like you think a bunch of cameras and computers would work? To me, it feels like an empirical problem that will take years of research.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T01:01:14.104Z · LW · GW

I'm not convinced about the difficulty of operationalizing Eliezer's doomer bet. Effectively, loaning money to a doomer who plans to spend it all by 2030 is, in essence, a claim on the doomer's post-2030 human capital. The doomer thinks it's worthless, whereas the skeptic thinks it has value. Hence, they transact.

The TAGI case seems trickier than the doomer case. Who knows what a one dollar bill will be worth in a post-TAGI world.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T00:54:14.325Z · LW · GW

Sounds good. Can also leave money out of it and put you down for 100 pride points. :)

If so, message me your email and I'll send you a calendar invite for a group reflection in 2043, along with a midpoint check in in 2033.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-09T00:50:13.909Z · LW · GW

Right, but what inputs and outputs would be sufficient to reward modeling of the real world? I think that might take some exploration and experimentation, and my 60% forecast is the odds of such inquiries succeeding by 2043.

Even with infinite compute, I think it's quite difficult to build something that generalizes well without overfitting.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-08T21:32:05.447Z · LW · GW

Gotcha. I guess there's a blurry line between program search and training. Somehow training feels reasonable to me, but something like searching over all possible programs feels unreasonable to me. I suppose the output of such a program search is what I might mean by an algorithm for AGI.

Hyperparameter search and RL on a huge neural net feels wildly underspecified to me. Like, what would be its inputs and outputs, even?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-08T21:15:05.749Z · LW · GW

Excellent comment - thanks for sticking your neck out to provide your own probabilities.

Given the gulf between our 0.4% and your 58.6%, would you be interested in making a bet (large or small) on TAI by 2043? If yes, happy to discuss how we might operationalize it.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T22:48:24.124Z · LW · GW

I'm curious and I wonder if I'm missing something that's obvious to others: What are the algorithms we already have for AGI? What makes you confident they will work before seeing any demonstration of AGI?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T08:04:28.446Z · LW · GW

If humans can teleoperate robots, why don't we have low-wage workers operating robots in high-wage countries? Feels like a win-win if the technology works, but I've seen zero evidence of it being close. Maybe Ugo is a point in favor?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T08:00:35.967Z · LW · GW

Interesting. When I participated in the AI Adversarial Collaboration Project, a study funded by Open Philanthropy and executed by the Forecasting Research Institute, I got the sense that most folks concerned about AI x-risk mostly believed that AGIs would kill us on their own accord (rather than by accident or as a result of human direction), that AGIs would have self-preservation goals, and therefore AGIs would likely only kill us after solving robotic supply chains (or enslaving/manipulating humans, as I argued as an alternative).

Sounds like your perception is that LessWrong folks don't think robotic supply chain automation will be a likely prerequisite to AI x-risk?

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T07:55:01.044Z · LW · GW

Yeah, that's a totally fair criticism. Maybe a better header would be "evidence of accuracy." Though even that is a stretch given we're only listing events in the numerators. Maybe "evidence we're not crackpots"?

Edit: Probably best would be "Forecasting track record." This is what I would have gone with if rewriting the piece today.

Edit 2: Updated the post.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T07:52:43.284Z · LW · GW

According to our rough and imperfect model, dropping inference needs by 2 OOMs increases our likelihood of hitting the $25/hr target by 20%abs, from 16% to 36%.

It doesn't necessarily make a huge difference to chip and power scaling, as in our model those are dominated by our training estimates, not our inference need estimates. (Though of course those figures will be connected in reality.)

With no adjustment to chip and power scaling, this yields a 0.9% likelihood of TAGI.

With a +15%abs bump to chip and power scaling, this yields a 1.2% likelihood of TAGI.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T07:44:47.836Z · LW · GW

Great points.

I think you've identified a good crux between us: I think GPT-4 is far from automating remote workers and you think it's close. If GPT-5/6 automate most remote work, that will be point in favor of your view, and if takes until GPT-8/9/10+, that will be a point in favor of mine. And if GPT gradually provides increasingly powerful tools that wildly transform jobs before they are eventually automated away by GPT-7, then we can call it a tie. :)

I also agree that the magic of GPT should update one into believing in shorter AGI timelines with lower compute requirements. And you're right, this framework anchored on the human brain can't cleanly adjust from such updates. We didn't want to overcomplicate our model, but perhaps we oversimplified here. (One defense is that the hugeness of our error bars mean that relatively large updates are needed to make a substantial difference in the CDF.)

Lastly, I think when we see GPT unexpectedly pass the Bar, LSAT, SAT, etc. but continue to fail at basic reasoning, it should update us into thinking AGI is sooner (vs a no pass scenario), but also update us into realizing these metrics might be further from AGI than we originally assumed based on human analogues.

Comment by Ted Sanders (ted-sanders) on Transformative AGI by 2043 is <1% likely · 2023-06-07T00:08:12.182Z · LW · GW

How has this forecast changed in the last 5 years?  Has widespread and rapid advance of non-transformative somewhat-general-purpose LLMs change any of your component predictions?

We didn't have this framework 5 years ago, but the tremendous success of LLMs can only be a big positive update, I think. That said, some negative updates for me from the past 15 years have been how slowly Siri improved, how slowly Wolfram Alpha improved, and how slowly Alexa improved. I genuinely expected faster progress from their data flywheels after their launches, but somehow it didn't seem to happen. Self-driving seems to be middle of the road compared to how I thought it would go 5 years ago.

I don't actually disagree, but MUCH of the cause of this is an excessively high bar (as you point out, but it still makes the title misleading). 

Agreed. I think the "<1%" headline feels like an aggressive claim, but the definition from the contest we use is a very high bar. For lower bars, we'd forecast much higher probabilities. We expect great things from AI and AGI, and we are not reflexively bearish on progress.