Some thoughts on automating alignment researchpost by Lukas Finnveden (Lanrian) · 2023-05-26T01:50:20.099Z · LW · GW · 3 comments
As AI systems get more capable, they may at some point be able to help us with alignment research. This increases the chance that things turn out ok. Right now, we don’t have any particularly scalable or competitive alignment solutions. But the methods we do have might let us use AI to vastly increase the amount of labor spent on the problem before AI has the capability and motivation to take over. In particular, if we’re only trying to make progress on alignment, the outer alignment problem is reduced to (i) recognising progress on sub-problems of alignment (potentially by imitating productive human researchers), and (ii) recognising dangerous actions like e.g. attempts at hacking the servers.
But worlds in which we’re hoping for significant automated progress on alignment are fundamentally scary. For one, we don’t know in what order we’ll get capabilities that help with alignment vs. dangerous capabilities. But even putting that aside, AIs will probably become helpful for alignment research around the same time as AIs become better at capabilities research. Once AI systems can significantly contribute to alignment (say, speed up research by >3x), superintelligence will be years or months away. (Though intentional, collective slow-downs could potentially buy us more time. Increasing the probability that such slow downs happen at key moments seems hugely important.)
In such a situation, we should be very uncertain about how things will go.
To illustrate one end of the spectrum: It’s possible that automated alignment research could save the day even in an extremely tense situation, where multiple actors (whether companies or nations) were racing towards superintelligence. I analyze this in some detail here. To briefly summarize:
- If a cautious coalition were to (i) competitively advance capabilities (without compromising safety) for long enough that their AI systems became really productive, and (ii) pause dangerous capabilities research at the right time — then even if they only had a fairly small initial lead, that could be enough to do a lot of alignment research.
- How could we get AI systems that significantly accelerates alignment research without themselves posing an unacceptable level of risk? It’s not clear that we will, but one possible story is motivated in this post [LW · GW]: It’s easier to align subhuman models than broadly superhuman models, and in the current paradigm, we will probably be able to run hundreds of thousands of subhuman models before we get broadly superhuman models, each of them thinking 10-100X faster than humans. Perhaps they could make rapid progress on alignment.
In a bit more detail:
Let’s say that a cautious coalition starts out with an X-month lead, meaning that it will take X months for other coalitions to catch up to their current level of technology. The cautious coalition can maintain that X-month lead for as long as they don't leak any technology, and for as long as they move as fast as their competitors.
In reality, the cautious coalition should gradually become more careful, which might gradually reduce their lead (if their competitors are less cautious). But as a simple model, let’s say that the cautious coalition maintains their X-month lead until further advancement would pose a significant takeover risk, at which point they entirely stop advancing capabilities and redirect all their effort towards alignment research.
Simplifying even further, let’s say that, when they pause, their current AI capabilities are such that 100 tokens of an end-to-end trained language model on average lead to equally much alignment progress as 1 human researcher second (and that it does so safely).
According to this other post [LW · GW], if that language model was trained on ¼ of today’s GPUs and TPUs running for a year, then:
You could get ~20 million tokens per second by running copies of that AI on its own training compute.
The post also guesses at a serial speed of ~50 tokens per second. Let’s assume that the previously-mentioned 100 tokens needed to replicate 1 human-researcher-second is 10 parallel streams of 10 tokens each.
Then each month of lead that the leader started out with would correspond to 15,000 human researchers working for 15 months.
Tricks like reduced precision or knowledge distillation might increase this further. And as described in the other post: If the language model was larger and trained on more compute, or if it was trained over a shorter period of time, then you would be able to run even more of them in parallel. For a model trained with 1000x the compute, over the course of 4 rather than 12 months, you could 100x as many models in parallel. You’d have 1.5 million researchers working for 15 months.
If the cautious coalition could keep increasing productivity without increasing the danger of their models (e.g. by making controllable models more efficient and running a greater number of them) then they could plausibly get even more. The doc linked above considers a model where this can yield 10-1000x more researcher hours.
If this worked out as described, this could lead to much progress in alignment. And plausibly also progress in other areas — e.g. demos of risk that could help persuade other companies or governments that a slow-down is needed.
To point out some significant issues with that particular analysis:
If the alignment methods crucially rely on human supervision, then research might be constrained by human supervisors — either by their lack of serial speed or by their small number. This could dramatically curtail the progress that AIs could make.
This technology will be very valuable, and so will be very tempting to steal. If the leader’s models are stolen, they would lose all of their lead at once, and their main effect on the world will have been to speed up how soon incautious actors can build superintelligence.
- This emphasizes the need for exceptional infosecurity in projects like this.
Even without theft, it’s not clear whether a project could afford to develop technology without deploying it, for long. Training runs will be expensive. Deployment is likely to leak information about how the technology is developed.
We may not get safe AI systems for which 100 tokens produce as much value as a human-researcher second. I more-or-less made up that number, and I would be interested in better analysis of how much useful work we could safely get with realistic alignment techniques. Three particular caveats are that:
The model of completely automating human researchers using N tokens is unrealistic, since AIs will have a very different distribution of talents than humans. Different tasks will be differently hard to automate.
There’s an important difference between what is safe and what we can be sufficiently confident is safe. The latter is a higher bar.
The amount of work that could be safely automated would depend on how prepared the cautious coalition was. These numbers aren’t exogenous — if you wanted this plan to work, you would want to do a lot of alignment research in advance.
Pausing capabilities work and/or redirecting a lot of effort to alignment could be a very large commitment in a very high-stakes situation. Therefore, it could require a large amount of institutional buy-in and/or up-front commitments and/or regulatory pressure.
Correctly implementing safe-guards, assessing risks along the way, and making good judgments about when to (not) press-on requires a high-bar for organizational adequacy [LW · GW]. It’s very easy to imagine a well-intentioned group of people catastrophically messing up the execution.
So reiterating: significant automation of alignment is a real possibility, which could save an otherwise hopeless situation. But worlds in which this needs to be done on a short deadline are very scary worlds — because it’s plausible that sufficient boosts won’t be possible to achieve without danger, or that the danger will be misjudged, directly leading to AI takeover. If we could collectively slow down, that would be much safer.
I’ll end this post by explicitly mentioning a few other strategic implications from the possibility of automating alignment research:
- It underscores the value of good governance, good coordination ability, and concern about the risks among labs and governments.
- These things are key for being able to slow down at the right time, which this perspective suggests could be extremely valuable. It’s possible that our ultimate understanding of alignment will depend less on how much alignment work we did ahead of time than on how willing and able key actors were to slow down and study the risks towards the end.
- If early AI systems are sufficiently good at all parts of alignment research, then they could end up obsoleting our efforts at technical research in a short amount of time. But coordination ability might crucially rest on precedent and agreements made beforehand.
- Alignment methods that aren’t arbitrarily scalable can still be really useful. If we can align systems of a given capability level, that can help us align the next capability level even if we wouldn’t have gotten there fast enough on our own.
- To be clear about terminology: This shifts me towards thinking that e.g. really well-executed debate could be great (despite not being able to get inaccessible information [LW · GW]). At current labs, debate is already termed “scalable” (since it’s importantly more scalable than unassisted RLHF).
- Alignment methods that aren’t very competitive can still be useful. If some internally-coordinated actor has a temporary lead, and is motivated by alignment concerns, they can potentially use uncompetitive methods to develop better alignment methods.
- For example: Spending a lot of human effort on checking output, using more process-based [LW · GW] feedback as opposed to outcomes-based feedback, etc.
- Though your methods can only be so uncompetitive before they make your AI so weak that AI doesn’t speed up alignment research much. So thinking about competitiveness is still important.
- This possibility isn’t captured in the toy model above. It’s assumed that the leader’s capabilities progress goes from maximally fast to a complete halt. A more realistic model could have the leader gradually make tradeoffs. (Some of this is captured in the doc mentioned above.)
- A notable chunk of the value of alignment research might come from learning enough about the problem that it’s easier to automate later on. If we get an opportunity to put vast numbers of subhuman models to work, it would be great to have well-scoped problems that they can work on.
- Though a big chunk also comes from worlds where (some types of) alignment research isn’t significantly sped up before we face large takeover risks. To reiterate: Counting on significant speed-ups is scary, and the less we need to rely on them, the better.
Acknowledgements: Thanks to Tom Davidson and Carl Shulman for comments and ideas. I work at Open Philanthropy but the views here are my own.
Though it’s closely related to the scary insight that AI might accelerate AI research, which suggests shorter timelines and faster takeoff speeds. If you had previously reflected on “AI accelerating capabilities”, but not on “AI accelerating alignment”, then the latter should come as a positive update. But if you were to simultaneously update on both of them, I don’t know what the net effect on p(doom) is. ↩︎
This wouldn’t solve inner alignment, by itself. But inner misalignment would probably be easier to notice in less capable models. And for models that aren’t yet superintelligent, you could reduce risk by reducing their opportunities to cause catastrophes and/or to escape the lab, including by training many AIs on recognising and sounding the alarm on such attempts. ↩︎
There are different kinds of dangerous capabilities. In particular, there’s a spectrum between (i) capabilities that are only dangerous if they’re developed and deployed by sufficiently incautious actors to (ii) capabilities that would make the models dangerous even if cautious actors combined them with the best available alignment techniques. The former end of the spectrum is important for the question “How long do we have until incautious actors cause AI takeover?” (absent successful efforts at regulating them or otherwise making them cautious). The latter is important for the question “What sort of models could a maximally cautious actor use without risking AI takeover?”, which is important for what speed-up to alignment research we could expect. ↩︎
Which highlights the strong importance of info security. Even with good info security, leaking 0 information might be unrealistic, if the leader needs to widely deploy their technology to finance training. ↩︎
More precisely: As long as they move as fast as their competitors will move when they get to the cautious coalitions’ level of AI technology. This might differ from their competitors’ current pace if novel AI systems can be used to accelerate capabilities research. So if nearby competitors will use AI to significantly accelerate their capabilities research, then this condition would require the cautious coalition to be the first to do so. At some point, that will probably become very dangerous. At that point, the cautious coalition would need to slow down or pause. ↩︎
According to this, there’s ~4e21 FLOP/s of computing capacity on GPUs and TPUs as of Q1 2023. (Most of this is on chips that would struggle to run SOTA models, so the numbers are more relevant for what you could get if chip manufacturers switched over to producing many more AI chips.) A quarter of this for a year would be 1e21*(365243600)*50% ~= 3e28. With 50% hardware utilization, this would correspond to a Chinchilla-trained model with 11 trillion parameters. ↩︎
Serial researcher months per month = (0.1 serial-researcher-second/serial-token) * (50 serial-tokens/second) * (36002430 seconds/month) / (3600830 serial-researcher-seconds/serial-researcher-months) = 15 serial researcher months / month.
In the above denominator, I use “8” instead of 24 hours per day, because humans only work ~⅓ of their life, while GPUs can run non-stop.
In total, there are (0.01 researcher-seconds / token) * (2.3e7 tokens/second) ~= 6e11 tokens per second. So given a serial speed-up of 15x, that corresponds to 6e11 / 15 ~= 15000 parallel researchers. ↩︎
A model with 1000x the compute (1.5e31) would need enough compute to process sqrt(1000)~=32x more data in parallel. And if it was trained in 12/4=3x shorter time, it would need 3x more compute in parallel. 32 * 3 ~= 100. ↩︎
Imagine that the AI researchers invent new programming languages and write huge, new code-bases in them. And then imagine a human trying to arbitrate a debate about whether a line in the code introduces a vulnerability. ↩︎
Among simplistic models you can use, it’s plausible to me that a better model would be to assume that tasks will be automated one at a time, and that once you’ve automated a task, it’s cheap to do it as much as you want (c.f. Tom Davidson’s takeoff speeds model [LW · GW] and Richard Ngo’s framework [LW · GW]). I’d be interested in more research on what this would suggest about alignment automation. ↩︎
This would be compatible with using AI assistance to make progress on key problems during the slow-down. If you had many eyes on that process and there was ample room to pause if something looked dangerous, then that would make AI assistance much safer than in a unilateral race-situation. ↩︎
Comments sorted by top scores.
comment by RogerDearnaley (roger-d-1) · 2023-05-28T09:26:28.575Z · LW(p) · GW(p)
As long as all the agentic AGIs people are building are value learners (i.e. their utility function is hard-coded to something like "figure out what utility function humans in aggregate would want you to use if they understood the problem better, and use that"), then improving their understanding of the human values becomes a convergent instrumental strategy for them: obviously the better they understand the human-desired utility function, the better job they can do of optimizing it. In particular, if AGI's capabilities are large, and as a result many of the things it can do are outside the region of validity of its initial model of human values, and also it understands the concept of the region of validity of a model (a rather basic, obviously required capability for an AGI that can do research, so this seems like a reasonable assumption), then it can't use most of its capabilities safely, so solving that problem obviously becomes top priority. This is painfully obvious to us, so it should also be painfully obvious to an AGI capable of doing research.
In that situation, a fast takeoff should just cause you to get an awful lot of AGI intelligence focused on the problem of solving alignment. So, as the author mentions, perhaps we should be thinking about how we would maintain human supervision in that eventuality? That strikes me as a particular problem that I'd feel more comfortable to have solved by a human alignment researcher than an AGI one.Replies from: alexander-gietelink-oldenziel
↑ comment by Alexander Gietelink Oldenziel (alexander-gietelink-oldenziel) · 2023-05-28T13:48:47.096Z · LW(p) · GW(p)
If we solve the alignment problem than we solve alignment problem.
I agree with this true statement.Replies from: roger-d-1
↑ comment by RogerDearnaley (roger-d-1) · 2023-05-29T04:48:11.642Z · LW(p) · GW(p)
If we can solve enough of the alignment problem, the rest gets solved for us.
If we can get a half-assed approximate solution to the alignment problem, sufficient to semi-align a STEM-capable AGI value learner of about smart-human level well enough to not kill everyone, then it will be strongly motivated to solve the rest of the alignment problem for us, just as the 'sharp left turn' is happening, especially if it's also going Foom. So with value learning, there is is a region of convergence around alignment.
Or to reuse one of Eliezer's metaphors, then if we can point the rocket on approximately the right trajectory, it will automatically lock on and course-correct from there.