post by [deleted] · · ? · GW · 0 comments

This is a link post for

0 comments

Comments sorted by top scores.

comment by Daniel Kokotajlo (daniel-kokotajlo) · 2021-04-23T06:19:10.372Z · LW(p) · GW(p)

Thanks for this post! I'll write a fuller response later, but for now I'll say: These arguments prove too much; you could apply them to pretty much any technology (e.g. self-driving cars, 3D printing, reusable rockets, smart phones, VR headsets...). There doesn't seem to be any justification for the 50-year number; it's not like you'd give the same number for those other techs, and you could have made exactly this argument about AI 40 years ago, which would lead to 10-year timelines now. You are just pointing out three reasons in favor of longer timelines and then concluding

it's a bit difficult to see how we will get transformative AI developments in the next 50 years. Even accepting some of the more optimistic assumptions in e.g. Ajeya Cotra's Draft report on AI timelines [LW · GW], it still seems to me that these effects will add a few decades to our timelines before things get really interesting.

Which seems unwarranted to me. I agree that the things you say push in the direction of longer timelines, but there are other arguments one could make that push in the direction of shorter timelines, and it's not like your arguments are so solid that we can just conclude directly from them that timelines are long--and specifically 50+ years long!

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2021-04-23T07:08:53.804Z · LW(p) · GW(p)

These arguments prove too much; you could apply them to pretty much any technology (e.g. self-driving cars, 3D printing, reusable rockets, smart phones, VR headsets...).

I suppose my argument has an implicit, "current forecasts are not taking these arguments into account." If people actually were taking my arguments into account, and still concluding that we should have short timelines, then this would make sense. But, I made these arguments because I haven't seen people talk about these considerations much. For example, I deliberately avoided the argument that according to the outside view, timelines might be expected to be long, since that's an argument I've already seen many people make, and therefore we can expect a lot of people to take it into account when they make forecasts.

I agree that the things you say push in the direction of longer timelines, but there are other arguments one could make that push in the direction of shorter timelines

Sure. I think my post is akin to someone arguing for a scientific theory. I'm just contributing some evidence in favor of the theory, not conducting a full analysis for and against it. Others can point to evidence against it, and overall we'll just have to sum over all these considerations to arrive at our answer.

Replies from: daniel-kokotajlo, None
comment by Daniel Kokotajlo (daniel-kokotajlo) · 2021-04-23T08:42:46.195Z · LW(p) · GW(p)

I definitely agree that our timelines forecasts should take into account the three phenomena you mention, and I also agree that e.g. Ajeya's doesn't talk about this much. I disagree that the effect size of these phenomena is enough to get us to 50 years rather than, say, +5 years to whatever our opinion sans these phenomena was. I also disagree that overall Ajeya's model is an underestimate of timelines, because while indeed the phenomena you mention should cause us to shade timelines upward, there is a long list of other phenomena I could mention which should cause us to shade timelines downward, and it's unclear which list is overall more powerful.

On a separate note, would you be interested in a call sometime to discuss timelines? I'd love to share my overall argument with you and hear your thoughts, and I'd love to hear your overall timelines model if you have one.

comment by [deleted] · 2021-04-24T07:14:59.937Z · LW(p) · GW(p)

Matthew, one general comment.  Most models of AI adoption once the conditions are reached are exponential.  So your forecast model is flawed in this way.

AI will take over in an area once it is : (robust (mostly software robustness), and it solves the general task with few edge cases).  Arguably it already has 'taken over' the space of board games in that if solving board games had economic value (in the way that loading a truck has economic value), all players would already be AI.

Once the conditions of (robustness, general task) is solved, or your arguments #1 and #3, there is a key fact you are missing:

          Regulatory agencies and people don't have a choice but to adopt.  That is, it's not a voluntary act.  They either do it or they go broke/cease to matter.  This is something I will break out just more generally: if a country has a (robust, general) AI agent that can drive cars, they can immediately save paying several million people.   This means that any nation that 'slows down' adoption via regulation becomes uncompetitive on the global scale, and any individual firm that 'slows down' adoption goes broke because it's competitors can sell services below marginal cost.  

Now, today there are problems.  We don't yet have a good framework to prove robustness.  It's actually a difficult software engineering task in itself.  It may ultimately turn out to be a harder problem than solving the general AI problem itself...*

Point is, you argument reduces to: "I believe it will take more than 50 years for a (robust, general) TAI to be developed where it exists in at least one place and is owned by an entity who intends to release it" 

And you might be right.  But it all hinges on your second argument.

 

*arguably the entire "alignment" problem is really a subset of "robustness".  

Replies from: logan-zoellner
comment by Logan Zoellner (logan-zoellner) · 2021-04-24T11:51:26.145Z · LW(p) · GW(p)

Regulatory agencies and people don't have a choice but to adopt.  That is, it's not a voluntary act.  They either do it or they go broke/cease to matter.  This is something I will break out just more generally: if a country has a (robust, general) AI agent that can drive cars, they can immediately save paying several million people.   This means that any nation that 'slows down' adoption via regulation becomes uncompetitive on the global scale, and any individual firm that 'slows down' adoption goes broke because it's competitors can sell services below marginal cost.  

 

This argument seems to prove too much.  If regulators absolutely cannot regulate something because they will get wiped out by competitors, why does overregulation exist in any domain?  Taking nuclear power as an example, it is almost certainly true that nuclear could be 10x cheaper than existing power sources with appropriate regulation, yet no country has done this.

The whole point is that regulators DO NOT respond to economic incentives because the incentives apply to those being regulated, not the regulator themselves.

Replies from: None
comment by [deleted] · 2021-04-24T12:07:01.461Z · LW(p) · GW(p)

Nuclear power is easily explained.  It doesn't fit the (robust, general) heuristic I mentioned above as it isn't robust.  Nor does it fit a third implied parameter, economic gain.  Implicitly a robust and general AI system provides economic gain because the cost of the compute electronics and the energy to run them is far less than the cost of upkeep of a human being.  (initially this would be true only in rich countries, but as the compute electronics become commodified it would soon be true almost everywhere)

Nuclear power, jetpacks, flying cars, Moon bases - most failed future predictions just fail the economic gain constraint.

Nuclear power is not 10x cheaper.  It carries large risks so some regulation cannot be skipped.  I concur that there is some unnecessary regulation, but the evidence such as the linked source just doesn't leave "room" for a 10x gain.  Currently the data suggests it doesn't provide an economic gain over natural gas unless the carbon emissions are priced in, and they are not in most countries.

The other items I mention also don't have an economic gain.  Jetpacks/flying cars - trivial, the value of the saved time is less than the value of the (fuel guzzled for a VTOL, capital cost/wear and tear on a personal VTOL, and externalities like noise and crashes).  For wealthy individuals where time is that valuable they do have VTOLs - helicopters - as they also value their continued existence, and a helicopter piloted by a professional is safer than a personal jetpack.

Moon base is similar, the scientific knowledge about a dead rock doesn't really "pay rent" sufficient to justify the cost of sending humans there.  

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2021-04-24T18:35:38.915Z · LW(p) · GW(p)

Nuclear power is not 10x cheaper.  It carries large risks so some regulation cannot be skipped.  I concur that there is some unnecessary regulation, but the evidence such as the linked source just doesn't leave "room" for a 10x gain.  Currently the data suggests it doesn't provide an economic gain over natural gas unless the carbon emissions are priced in, and they are not in most countries.

I recommend reading the Roots of Progress article I linked to in the post. Most of the reason why nuclear power is high cost is because of the burdensome regulations. And of course, regulation is not uniformly bad, but it seems from the chart Devanney Figure 7.11 in the article that we could have relatively safe nuclear energy for a fraction of its current price.

Replies from: None
comment by [deleted] · 2021-04-24T19:48:46.764Z · LW(p) · GW(p)

Ok, I look at the chart.  It seems to show that in the 1970s the cost per kw of capacity of nuclear hit a trough at about $1 a watt*.  Was this corrected for inflation?  And that the cost of new capacity has soured to ridiculous levels.

We still have many of those reactors built in the 1970s.  They are linked in the lazard data above as 'paid for' reactors.  They are $29 a megawatt-hour.  Solar hits as low as $31 a megawatt-hour, and natural gas $28 in the same 'paid for case'.

So it appears that no, actually, we cannot get energy for 10x lower than the current price.  (I think as a rational agent you need to 'update' now or prove that this statement is false?)

My other point was that if other nations could do it - not all of them have the same regulatory scheme.  If other nations could build reactors at a fraction of the price they would benefit.  And China has a strong incentive if this were true - they have a major pollution problem with coal.  But, "the industry has not broken ground on a new plant in China since late 2016".  

So this suggests that either : the inefficient and unproductive regulatory scheme you mention is so 'viral' that even a country that appears to be able to shove through other major changes overnight just can't help itself but to make nuclear unproductive through too much regulation.  Or in fact there isn't the opportunity for 10x lower costs, that in fact a nuclear reactor is complicated and made of expensive parts that are made in tiny quantities and hard to reduce in cost, that after a reactor is paid for it still requires a huge crew to care and feed it, and that the radiation fields whenever there is a leak or need to work on certain areas of the plant make it far more difficult and expensive to work on, even though the risk to the general public may be very low.  Oh and the government has an incentive to carefully monitor every nuclear reactor just to make sure someone isn't making plutonium.  

Back to the original topic of AI timelines: with AI systems there isn't a 10 year investment or a need to get specialized parts only made in Japan to make an AI. It's not just the software you will need, AI systems do need specialized compute platforms, but there are multiple vendors for these and most countries are going to be able to buy as much of them as they want.  Therefore if a country can be lax in regulating AI systems and get "10x lower labor costs" by having the AI systems do labor instead of humans, they get an economic benefit.  Therefore unless regulatory regimes are so "viral" they take over everywhere, in every nation, and prevent this everywhere, you see AI grow like wildfire in certain places, and everyone else will be forced to laxen their rules or be left behind.

As a simple example, if China banned near-term AI systems or regulated them too severely, but Australia allowed them, Australia could use them to mine for resources in deep mines too dangerous for humans and then build self replicating factories.  In 5-10 years of exponential growth their industrial output would exceed all of China's with 25 million people + a few million AI specialists they might need to bring in if they can't work remotely.  [due to regulations]

So either China has to allow them to 'keep up' or stop being a superpower.  

Replies from: logan-zoellner
comment by Logan Zoellner (logan-zoellner) · 2021-05-01T05:04:57.476Z · LW(p) · GW(p)

We still have many of those reactors built in the 1970s.  They are linked in the lazard data above as 'paid for' reactors.  They are $29 a megawatt-hour.  Solar hits as low as $31 a megawatt-hour, and natural gas $28 in the same 'paid for case'.

 

Your claim here is that under optimal regulatory policy we could not possibly do better today than with 1970's technology?

 

My other point was that if other nations could do it - not all of them have the same regulatory scheme.  If other nations could build reactors at a fraction of the price they would benefit.  And China has a strong incentive if this were true - they have a major pollution problem with coal.  But, "the industry has not broken ground on a new plant in China since late 2016".  

from the article you linked

The 2011 meltdown at Japan’s Fukushima Daiichi plant shocked Chinese officials and made a strong impression on many Chinese citizens. A government survey in August 2017 found that only 40% of the public supported nuclear power development.

It seems perfectly reasonable to believe China too can suffer from regulatory failure due to public misconception.   In fact, given it's state-driven economy, wouldn't we expect market forces to be even less effective at finding low-cost solutions than in Western countries?  Malinvestment seems to be a hallmark of the current Chinese system.

Replies from: None
comment by [deleted] · 2021-05-01T18:08:46.726Z · LW(p) · GW(p)

Your claim here is that under optimal regulatory policy we could not possibly do better today than with 1970's technology?

Yes.  I do claim that.  Even if the reactors were 'free' they are still not better than solar/wind.  So if the regulatory agencies decided to raise the accepted radiation doses by many orders of magnitude, and to just stop requiring any protections at all - ok to build a nuclear reactor in a warehouse - I am saying it wouldn't be cost effective.

If only we had a real world example of such a regime.  Oh wait, we do.  

comment by Steven Byrnes (steve2152) · 2021-04-22T19:22:01.167Z · LW(p) · GW(p)

Thanks for the nice post! Here's why I disagree :)

Technological deployment lag

Normal technologies require (1) people who know how to use the technology, and (2) people who decide to use the technology. If we're thinking about a "real-deal AGI" that can do pretty much every aspect of a human job but better and cheaper, then (1) isn't an issue because the AGI can jump into existing human roles. It would be less like "technology deployment" and more like a highly-educated exquisitely-skilled immigrant arriving into a labor market. Such a person would have no trouble getting a job, in any of a million different roles, in weeks not decades. For (2), the same "real-deal AGI" would be able to start companies of its own accord, build factories, market products and services, make money, invest it in starting more companies, etc. etc. So it doesn't need anyone to "decide to use the technology" or to invest in the technology.

Regulation will slow things down

I think my main disagreement comes from my thinking of AGI development as being "mostly writing and testing code inside R&D departments", rather than "mostly deploying code to the public and learning from that experience". I agree that it's feasible and likely for the latter activity to get slowed down by regulation, but the former seems much harder to regulate for both political reasons and technical reasons.

The political is: It's easy to get politicians riled up about the algorithms that Facebook is actually using to influence people, and much harder to get politicians riled up about whatever algorithms Facebook is tinkering with (but not actually deploying) in some office building somewhere. I think there would only be political will once we start getting "lab escape accidents" with out-of-control AGIs self-replicating around the internet, or whatever, at which point it may well be too late already.

The technical is: A lot of this development will involve things like open-source frameworks to easily parallelize software, and easier-to-use faster open-source implementations of new algorithms, academic groups publishing papers, and so on. I don't see any precedent or feasible path for the regulation of these kinds of activities, even if there were the political will.

Not that we shouldn't develop political and technical methods to regulate that kind of thing—it seems like worth trying to figure out—just that it seems extremely hard to do and unlikely to happen.

Overestimating the generality of AI technology

My own inside-view story (see here for example [LW · GW]) is that human intelligence is based around a legible learning algorithm, and that researchers in neuroscience and AI are making good progress in working out exactly how that learning algorithm works, especially in the past 5 years. I'm not going to try to sell you on that story here, but fwiw it's a short-ish timelines story that doesn't directly rely on the belief that currently-popular deep learning models are very general, or even necessarily on the right track.

Replies from: Jack Ryan
comment by Jack R (Jack Ryan) · 2021-04-24T04:42:47.377Z · LW(p) · GW(p)

Won't we have AGI that is slightly less able to jump into existing human roles before we have AGI that can jump into existing human roles? (Borrowing intuitions from Christiano's Takeoff Speeds) [Edited to remove typo]

Replies from: steve2152, None
comment by Steven Byrnes (steve2152) · 2021-04-24T12:15:49.136Z · LW(p) · GW(p)

Sure but that would make OP's point weaker not stronger, right?

comment by [deleted] · 2021-04-24T06:58:33.712Z · LW(p) · GW(p)

Jack, to be specific, we expect to have AI that can jump into specific classes of roles, and take over the entire niche.  All of it.  They will be narrowly superhuman at any role inside the class.  

If right now, strategy games, both of the board and the realtime clicking variety, had direct economic value, every human doing it would already be superfluous.  We can fully solve the entire class.  The reason is, succinctly:

a.  Every game-state can be modeled on a computer with the subsequent state resulting from a move by the AI agent provided

b.  The game-state can be reliably converted to a score that is an accurate assessment of what we care about - victory in the game.  That is, it's usually a delayed reward, but a game-state either is winning or it is not and this mapping is reliable.

For real world tasks (b) gets harder because there are subtle outcomes that can't be immediately perceived, or they are complex to model.  Example: an autonomous car reaches the destination but has damaged it's own components more than the value of the ride.  

So it will take longer to solve the class of :

          robotics manipulation problems where we can reliably estimate the score resulting from a manipulation, and model reasonably accurately the full environment and the machine in that environment.

This is most industrial and labor tasks on the planet in this class.  But the whole class can be solved relatively quickly - once you have a general solver for part of it, the rest of it will fall.

And then the next class of tasks are things where a human being is involved.  Humans are complex and we can't model them in a simulator like we can model rigid bodies and other physics.  I can't predict when this class will be solved.

comment by Donald Hobson (donald-hobson) · 2021-04-23T09:55:34.406Z · LW(p) · GW(p)

I don't think technological deployment is likely to take that long for AI's. With a physical device like a car or fridge, it takes time for people to set up the factories, and manufacture the devices. AI can be sent across the internet in moments. I don't know how long it takes google to go from say an algorithm that detects streets in satellite images to the results showing up in google maps, but its not anything like the decades it took those physical techs to roll out.

The slow roll-out scenario looks like this, AGI is developed using a technique that fundamentally relies on imitating humans, and requires lots of training data. There aren't nearly enough data from humans that are AI experts to make an AI AI expert. The AI is about as good at AI research as the median human. Or maybe the 80th percentile human. Ie no good at all. The AI design fundamentally requires custom hardware to run at reasonable speeds. Add in some political squabbling and it could take a fair few years before wide use, although there would still be huge economic incentive to create it. 

The fast scenario is the rapidly self improving superintelligence. Where we have oodles of compute by the time we crack the algorithms. All the self improvement happens very fast in software. Then the AI takes over the world. (I question that "a few weeks" is the fastest possible timescale for this. )

(For that matter, the curves on the right of the graph look steeper. It takes less time for an invention to be rolled out nowadays)

For your second point, you can name biases that might make people underestimate timelines, I can name biases that might make people overestimate timelines. (eg Failure to consider techniques not known to you) And it all turns into a bias naming competition. Which is hardly truth tracking at all.

As for regulation, I think its what people are doing in R&D labs, not what is rolled out that matters. And that is harder to regulate. I also explicitly don't expect any AI Chernobyl. I don't strongly predict there won't be an AI Chernobyl either. I feel that if the relevant parties act with the barest modicum of competence,  there won't be an AI Chernobyl. And the people being massively stupid will carry on being massively stupid after any AI Chernobyl.

Replies from: logan-zoellner
comment by Logan Zoellner (logan-zoellner) · 2021-04-24T11:55:28.012Z · LW(p) · GW(p)

I don't think technological deployment is likely to take that long for AI's. With a physical device like a car or fridge, it takes time for people to set up the factories, and manufacture the devices. AI can be sent across the internet in moments.

 

Most economically important uses of AGI (self driving cars, replacing fast-food workers) require physical infrastructure.  There are some areas (e.g. high frequency stock trading and phone voice assistants) that do not, but those are largely automated already so there won't be a sudden boost when the "cross the threshold" of AGI.

Replies from: donald-hobson
comment by Donald Hobson (donald-hobson) · 2021-04-24T17:46:15.019Z · LW(p) · GW(p)

Surely the set of jobs an AGI could do out of the box is wider than that. Lets compare it to the set of jobs that can be done from home over the internet. Most jobs that can be done over the internet can be done by the AI. Judging by how much working from home has been a thing recently, a significant percentage of the economy. Plus a whole load of other jobs that only make sense when the cost of labour is really low, and or the labour is really fast. And I would expect the amount to increase with robotisation. (If you take an existing robot, and put an AGI on it, suddenly it can do a lot more useful stuff.)

Replies from: logan-zoellner
comment by Logan Zoellner (logan-zoellner) · 2021-05-01T04:35:39.052Z · LW(p) · GW(p)

In 2020 the average number of days that Americans teleworked doubled from 2.4 to 5.8 per month.  If we assume that 100% of that work could be done by AGI and that all of those working days were replaced in a single year, that would be a 29% boost to productivity, just barely above the 25%/year growth definition of TAI.

It is unlikely that 100% of such work can be automated (for example at-home learning makes up a large fraction of telework).  And much of what can be automated will be automated long before we reach AGI (travel agents, real estate, ...).

I'm not sure how putting AGI on existing robots makes them automatically more useful?  Neither my roomba nor car manufacturing robots (to pick two extremes) can be greatly improved by additional intelligence.  Undoubtedly self-driving cars would be much easier (perhaps trival) to implement given AGI, but self-driving cars are almost certainly a less than AGI-hard task.  Did you have some particular examples in mind of existing robots that need/benefit from AGI specifically?

comment by Jack R (Jack Ryan) · 2021-04-24T03:34:39.040Z · LW(p) · GW(p)

Re: 1, I think it may be important to note that adoption has gotten quicker (e.g. as visualized in Figure 1 here; linking this instead of the original source since you might find other parts of the article interesting). Does this update you, or were you already taking this into account? 

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2021-04-24T18:39:02.023Z · LW(p) · GW(p)

Wow, that chart definitely surprised me. Yes, this caused me to update.

comment by Rohin Shah (rohinmshah) · 2021-04-23T19:31:44.494Z · LW(p) · GW(p)

I broadly agree with these points, and (1) and (3) in particular lead to me to shade the bio anchors estimates upwards by ~5 years (note they are already shaded up somewhat to account for these kinds of effects).

I don't really agree on (2).

I see no strong reason to doubt the narrow version of this thesis. I believe it's likely that, as training scales, we'll progressively see more general and more capable machine learning models that can do a ton of impressive things, both on the stuff we expect them to do well on, and some stuff we didn't expect.

But no matter how hard I try, I don't see any current way of making some descendant of GPT-3, for instance, manage a corporation.

I feel like if you were applying this argument to evolution, you'd conclude that humans would be unable to manage corporations, which seems too much. Humans seem to do things that weren't in the ancestral environment, why not GPTs, for the same reason?

You might say "okay, sure, at some level of scaling GPTs learn enough general reasoning that they can manage a corporation, but there's no reason to believe it's near". But one of the major points of the bio anchors framework is to give a reasonable answer to the question of "at what level of scaling might this work", so I don't think you can argue that current forecasts are ignoring (2).

Perhaps you just mean that most people aren't taking bio anchors into account and that's why (2) applies to them -- that seems plausible, I don't have strong beliefs about what other people are thinking.

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2021-04-23T20:16:13.249Z · LW(p) · GW(p)

Thanks for the useful comment.

You might say "okay, sure, at some level of scaling GPTs learn enough general reasoning that they can manage a corporation, but there's no reason to believe it's near".

Right. This is essentially the same way we might reply to Claude Shannon if he said that some level of brute-force search would solve the problem of natural language translation.

one of the major points of the bio anchors framework is to give a reasonable answer to the question of "at what level of scaling might this work", so I don't think you can argue that current forecasts are ignoring (2).

Figuring out how to make a model manage a corporation involves a lot more than scaling a model until it has the requisite general intelligence to do it in principle if its motivation were aligned.

I think it will be hard to figure out how to actually make models do stuff we want. Insofar as this is simply a restatement of the alignment problem, I think this assumption will be fairly uncontroversial around here. Yet, it's also a reason to assume that we won't simply obtain transformative models the moment they become theoretically attainable.

It might seem unfair that I'm inputting safety and control as an input in our model for timelines, if we're using the model to reason about the optimal time to intervene. But I think on an individual level it makes sense to just try to forecast what will actually happen.

Replies from: rohinmshah
comment by Rohin Shah (rohinmshah) · 2021-04-23T22:00:27.325Z · LW(p) · GW(p)

I think it will be hard to figure out how to actually make models do stuff we want. Insofar as this is simply a restatement of the alignment problem, I think this assumption will be fairly uncontroversial around here.

Fwiw, the problem I think is hard is "how to make models do stuff that is actually what we want, rather than only seeming like what we want, or only initially what we want until the model does something completely different like taking over the world".

I don't expect that it will be hard to get models that look like they're doing roughly the thing we want; see e.g. the relative ease of prompt engineering or learning from human preferences. If I thought that were hard, I would agree with you.

I would guess that this is relatively uncontroversial as a view within this field? Not sure though.

(One of my initial critiques of bio anchors was that it didn't take into account the cost of human feedback, except then I actually ran some back-of-the-envelope calculations and it turned out it was dwarfed by the cost of compute; maybe that's your crux too?)

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2023-01-30T22:24:04.778Z · LW(p) · GW(p)

Sorry for replying to this comment 2 years late, but I wanted to discuss this part of your reasoning,

Fwiw, the problem I think is hard is "how to make models do stuff that is actually what we want, rather than only seeming like what we want, or only initially what we want until the model does something completely different like taking over the world".

I think that's what I meant when I said "I think it will be hard to figure out how to actually make models do stuff we want". But more importantly, I think that's how most people will in fact perceive what it means to get a model to "do what we want".

Put another way, I don't think people will actually start using AI CEOs just because we have a language model that acts like a CEO. Large corporations will likely wait until they're very confident in its reliability, robustness, and alignment. (Although idk, maybe some eccentric investors will find the idea interesting, I just expect that most people will be highly skeptical without strong evidence that it's actually better than a human.)

I think this point can be seen pretty easily in discussion of driverless cars. Regulators are quite skeptical of Tesla's autopilot despite it seeming to do what we want in perhaps over 99% of situations. 

If anything, I expect most people to be intuitively skeptical that AI is really "doing what we want" even in cases where it's genuinely doing a better job than humans, and doesn't merely appear that way on the surface. The reason is simple: we have vast amounts of informal data on the reliability of humans, but very little idea how reliable AI will be. That plausibly causes people to start with a skeptical outlook, and only accept AI in safety-critical domains when they've seen it accumulate a long track record of exceptional performance.

For these reasons, I don't fully agree that "one of the major points of the bio anchors framework is to give a reasonable answer to the question of "at what level of scaling might this work"". I mean, I agree that this was what the report was trying to answer, but I disagree that it answered the question of when we will accept and adopt AI for various crucial economic activities, even if such systems were capable of automating everything in principle.

Replies from: rohinmshah
comment by Rohin Shah (rohinmshah) · 2023-02-02T09:36:12.683Z · LW(p) · GW(p)

I want to distinguish between two questions:

  1. At some specified point in the future, will people believe that AI CEOs can perform the CEO task as well as human CEOs if deployed?
  2. At some specified point in the future, will AI CEOs be able to perform the CEO task as well as human CEOs if deployed?

(The key difference being that (1) is a statement about people's beliefs about reality, while (2) is a statement about reality directly.)

(For all of this I'm assuming that an AI CEO that does the job of CEO well until the point that it executes a treacherous turn counts as "performing the CEO task well".)

I'm very sympathetic to skepticism about question 1 on short timelines, and indeed as I mentioned I agree with your points (1) and (3) in the OP and they cause me to lengthen my timelines for TAI relative to bio anchors.

My understanding was that you are also skeptical about question 2 on short timelines, and that was what you were arguing with your point (2) on overestimating generality. That's the part I disagree with. But your response is talking about things that other people will believe, rather than about reality; I already agree with you on that part.

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2023-02-02T19:14:15.555Z · LW(p) · GW(p)

I think I understand my confusion, at least a bit better than before. Here's how I'd summarize what happened.

I had three arguments in this essay, which I thought of as roughly having the following form:

  1. Deployment lag: after TAI is fully developed, how long will it take to become widely impactful?
  2. Generality: how difficult is it to develop TAI fully, including making it robustly and reliably achieve what we want?
  3. Regulation: how much will people's reactions to and concerns about AI delay the arrival of fully developed TAI?

You said that (2) was already answered by the bio anchors model. I responded that bio anchors neglected how difficult it will be to develop AI safely. You replied that it will be easy make models to seemingly do what we want, but that the harder part will be making models that actually do what we want.

My reply was trying to say that the inherent difficulty of building TAI safely was inherently baked into (2) already. That might be a dubious reading of the actual textual argument for (2), but I think that interpretation is backed up by my initial reply to your comment.

The reason why I framed my later reply as being about perceptions was because I think the requisite capability level at which people begin to adopt TAI is an important point about how long timelines will be independent of (1) and (3). In other words, I was arguing that people's perceptions of the capability of AI will cause them wait to adopt AI until it's fully developed in the sense I described above; it won't just delay the effects of TAI after it's fully developed, or before then because of regulation.

Furthermore, I assumed that you were arguing something along the lines of "people will adopt AI once it's capable of only seeming to do what we want", which I'm skeptical of. Hence my reply to you.

My understanding was that you are also skeptical about question 2 on short timelines, and that was what you were arguing with your point (2) on overestimating generality.

Since for point 2 you said "I'm assuming that an AI CEO that does the job of CEO well until the point that it executes a treacherous turn", I am not very skeptical of that right now. I think we could probably have AIs do something that looks very similar to what a CEO would do within, idk, maybe five years.

(Independently of all of this, I've updated towards medium rather than long timelines in the last two years, but mostly because of reflection on other questions, and because I was surprised by the rate of recent progress, rather than because I have fundamental doubts about the arguments I made here, especially (3), which I think is still underrated. 

ETA: though also, if I wrote this essay today I would likely fully re-write section (2), since after re-reading it I now don't agree with some of the things I said in it. Sorry if I was being misleading by downplaying how poor some of those points were.)

Replies from: rohinmshah
comment by Rohin Shah (rohinmshah) · 2023-02-04T08:57:07.128Z · LW(p) · GW(p)

My summary of your argument now would be:

  1. Deployment lag: it takes time to deploy stuff
  2. Worries about AI misalignment: the world will believe that AI alignment is hard, and so avoid deploying it until doing a lot of work to be confident in alignment.
  3. Regulation: it takes time to comply with regulations

If that's right, I broadly agree with all of these points :)

(I previously thought you were saying something very different with (2), since the text in the OP seems pretty different.)

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2023-02-04T11:37:19.702Z · LW(p) · GW(p)

I previously thought you were saying something very different with (2), since the text in the OP seems pretty different.

FWIW I don't think you're getting things wrong here. I also have simply changed some of my views in the meantime.

That said, I think what I was trying to accomplish with (2) was not that alignment would be hard per se, but that it would be hard to get an AI to do very high-skill tasks in general, which included aligning the model, since otherwise it's not really "doing the task" (though as I said, I don't currently stand by what I wrote in the OP, as-is).

comment by Rohin Shah (rohinmshah) · 2021-04-23T22:40:56.909Z · LW(p) · GW(p)

Planned summary for the Alignment Newsletter:

This post outlines and argues for three reasons to expect long AI timelines that the author expects are not taken into account in current forecasting efforts:

1. **Technological deployment lag:** Most technologies take decades between when they're first developed and when they become widely impactful.
2. **Overestimating the generality of AI technology:** Just as people in the 1950s and 1960s overestimated the impact of solving chess, it seems likely that current people are overestimating the impact of recent progress, and how far it can scale in the future.
3. **Regulation will slow things down,** as with [nuclear energy](https://rootsofprogress.org/devanney-on-the-nuclear-flop), for example.

You might argue that the first and third points don’t matter, since what we care about is when AGI is _developed_, as opposed to when it becomes widely deployed. However, it seems that we continue to have the opportunity to intervene until the technology becomes widely impactful, and that seems to be the relevant quantity for decision-making. You could have some specific argument like “the AI goes FOOM and very quickly achieves all of its goals” that then implies that the development time is the right thing to forecast, but none of these seem all that obvious.

Planned opinion:

I broadly agree that (1) and (3) don’t seem to be discussed much during forecasting, despite being quite important. (Though see e.g. [value of the long tail](https://www.lesswrong.com/posts/Nbcs5Fe2cxQuzje4K/value-of-the-long-tail).) I disagree with (2): while it is obviously possible that people are overestimating recent progress, or overconfident about how useful scaling will be, there has at least been a lot of thought put into that particular question -- it seems like one of the central questions tackled by <@bio anchors@>(@Draft report on AI timelines@). See more discussion in this [comment thread](https://www.alignmentforum.org/posts/Z5gPrKTR2oDmm6fqJ/three-reasons-to-expect-long-ai-timelines?commentId=F7FNee8Bpa8hemQkd).

comment by Tamsin Leake (carado-1) · 2023-01-31T05:53:15.989Z · LW(p) · GW(p)

i think we die probly this decade, and if not then probly next decade.

i partly explain my short timelines here. the short version is that i think recursive self-improvement (RSI) is not very hard to build. to kill everyone, you don't need lots of compute (though it helps), which means you don't need to be in a lab, which means you're not affected by regulation unless the regulation is "all computers are banned", which it's not going to be. you don't need to build "general" AI (whatever that means), you just need to build RSI. the hardest variable to predict is how many people are gonna be trying to build something like RSI, which is why my prediction is as vague as it is, but i think doom from RSI could happen any day, it just gets more likely on future days than on, say, tomorrow, because more people are trying with more powerful computers and more powerful available AI tech they can use.

comment by [deleted] · 2021-04-25T10:11:00.176Z · LW(p) · GW(p)

I agree that 1-3 need more attention, thanks for raising them.

Many AI scientists in the 1950s and 1960s incorrectly expected that cracking computer chess would automatically crack other tasks as well.

There’s a simple disconnect here between chess and self-supervised learning.  You're probably aware of it but it it's worth mentioning. Chess algorithms were historically designed to win at chess. In contrast, the point of self-supervised learning is to extract representations that are useful in general. For example, to solve a new tasks we can feed the representations into a linear regression, another general algorithm. ML researchers have argued for ages that this should work and we already have plenty of evidence that it does.

comment by Charlie Steiner · 2021-04-23T03:53:25.303Z · LW(p) · GW(p)

I'm curious how you'd estimate the relative importance of these factors. Myself, I think one of them really outweighs the others.

Replies from: matthew-barnett
comment by Matthew Barnett (matthew-barnett) · 2021-04-23T04:38:57.662Z · LW(p) · GW(p)

I'm uncertain. I lean towards the order I've written them in as the order of relative importance. However, the regulation thing seems like the biggest uncertainty to me. I don't feel like I'm good at predicting how people and government will react to things; it's possible that technological advancement will occur so rapidly and will be celebrated so widely that people won't want it to stop.

Replies from: donald-hobson
comment by Donald Hobson (donald-hobson) · 2021-04-23T14:25:14.733Z · LW(p) · GW(p)

Its quite possible governments don't really notice arriving AGI until its already there. Especially if the route taken is full of dense technical papers, with not much to impress the nonexpert. 

Its also possible that governments want to stop development, but find they basically can't. Ban AI research and everyone just changes the title from "AI" to "maths" or "programming" and does the same research. 

Replies from: None
comment by [deleted] · 2021-04-24T07:05:52.263Z · LW(p) · GW(p)

Note also that this technology, by it's definition ('transformative', 'intelligence') is so valuable it immediately gives whoever has it an absurd economic, military, even cultural advantage.  

Hard to compete with a country that can sell exports below your marginal cost, and of maybe knock off products that have better designs than the product they knocked off.  (the low cost is from self replicating robots, the reason the knock off design is better is an AI agent modeled the product being used and explored a large number of possible designs until it found a more reliable/cheaper to make one with similar functionality)

Military is because weapons are mostly a quantity thing, and self replicating robots can't really be beat there.

And cultural - as you have seen, a lot of cool tricks are possible with various generative AIs.  Presumably a 'transformative' one could do even cooler tricks.