mattmacdermott's Shortform
post by mattmacdermott · 2024-01-03T09:08:14.015Z · LW · GW · 19 commentsContents
19 comments
19 comments
Comments sorted by top scores.
comment by mattmacdermott · 2024-12-20T08:48:25.962Z · LW(p) · GW(p)
Could mech interp ever be as good as chain of thought?
Suppose there is 10 years of monumental progress in mechanistic interpretability. We can roughly - not exactly - explain any output that comes out of a neural network. We can do experiments where we put our AIs in interesting situations and make a very good guess at their hidden reasons for doing the things they do.
Doesn't this sound a bit like where we currently are with models that operate with a hidden chain of thought? If you don't think that an AGI built with the current fingers-crossed-its-faithful paradigm would be safe, what percentage an outcome would mech interp have to hit to beat that?
Seems like 99+ to me.
Replies from: Seth Herd, dtch1997↑ comment by Seth Herd · 2024-12-21T02:56:46.259Z · LW(p) · GW(p)
I very much agree. Do we really think we're going to track a human-level AGI let alone a superintelligence's every thought, and do it in ways it can't dodge if it decides to?
I strongly support mechinterp as a lie detector, and it would be nice to have more as long as we don't use that and control methods to replace actual alignment work and careful thinking. The amount of effort going into interp relative to the theory of impact seems a bit strange to me.
↑ comment by Daniel Tan (dtch1997) · 2024-12-21T23:08:21.349Z · LW(p) · GW(p)
Similar point is made here (towards the end): https://www.lesswrong.com/posts/HQyWGE2BummDCc2Cx/the-case-for-cot-unfaithfulness-is-overstated [LW · GW]
I think I agree, more or less. One caveat is that I expect RL fine-tuning to degrade the signal / faithfulness / what-have-you in the chain of thought, whereas the same is likely not true of mech interp.
comment by mattmacdermott · 2024-01-18T08:09:51.975Z · LW(p) · GW(p)
LW feature I would like: I click a button on a sequence and recieve one post in my email inbox per day.
comment by mattmacdermott · 2024-12-05T10:02:57.771Z · LW(p) · GW(p)
I'm confused by this argument from Redwood's AI control paper.
5.2 WHY WE DIDN’T CONSIDER STRATEGIES THAT INVOLVE TRAINING THE MODEL [...] Any protocol that involves training the model can be transformed into a validation-style protocol that merely validates the model’s performance on the dataset that was used as a train set. In any case where the validation-style protocol fails, we know that the red-team model is not distinguished from an aligned model by the measurements done by the protocol. Given that none of the training examples allow the reward-generating process to determine whether the policy it is training is undesirable, we don’t see why training would transform the model from an undesirable policy to a desirable policy. It is conceivable that we could find a training scheme that systematically converts undesirable policies to desirable ones, despite the lack of a reward differential causing it to do so. But we don’t know how.
If I understand correctly, the argument is:
- In order for training on dataset D to convince you a model is desirable, good performance on D needs to be a sufficient condition for being desirable.
- If good performance on D is a sufficient condition for being desirable, you can just test performance on D to decide whether a model is desirable.
But 2 seems wrong -- if good performance on D is sufficient but not neccesary for being desirable, then you could have two models, one desirable and one undesirable, which are indistinguishable on D because they both perform badly, and then turn them both into desirable models by training on D.
As an extreme example, suppose your desirable model outputs correct code and your undesirable model backdoored code. Using D you train them both to output machine checkable proofs of the correctness of the code they write. Now both models are desirable because you can just check correctness before deploying the code.
comment by mattmacdermott · 2024-01-03T09:08:14.193Z · LW(p) · GW(p)
I feel like there's a bit of a motte and bailey in AI risk discussion, where the bailey is "building safe non-agentic AI is difficult" and the motte is "somebody else will surely build agentic AI".
Are there any really compelling arguments for the bailey? If not then I think "build an oracle and ask it how to avoid risk from other people building agents" is an excellent alignment plan.
Replies from: quetzal_rainbow↑ comment by quetzal_rainbow · 2024-01-03T10:04:02.177Z · LW(p) · GW(p)
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-01-03T10:31:46.941Z · LW(p) · GW(p)
AIs limited to pure computation (Tool AIs) supporting humans, will be less intelligent, efficient, and economically valuable than more autonomous reinforcement-learning AIs (Agent AIs) who act on their own and meta-learn, because all problems are reinforcement-learning problems.
Isn’t this a central example of “somebody else will surely build agentic AI?”.
I guess it argues “building safe non-agentic AI before somebody else builds agentic AI is difficult” because agents have a capability advantage.
This may well be true (but also perhaps not, because e.g. agents might have capability disadvantages from misalignment, or because reinforcement learning is just harder than other forms of ML).
But either way I think it has importantly different strategy implications to “it seems difficult to make non-agentic AI safe”.
Replies from: quetzal_rainbow↑ comment by quetzal_rainbow · 2024-01-03T11:04:27.977Z · LW(p) · GW(p)
Oh, sorry, I misread your post.
I think the main problem with building safe non-agentic AI is that we don't know exactly what to build. It's easy to imagine how you type question in terminal, get an answer and then live happily ever after. It's hard to imagine what internals your algorithm should have to display this behaviour.
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-01-03T11:19:29.106Z · LW(p) · GW(p)
I think the most obvious route to building an oracle is to combine a massive self-supervised predictive model with a question-answering head.
What’s still difficult here is getting a training signal that incentives truthfulness rather than sycophancy, which is I think is what ARC‘s ELK stuff wants (wanted?) to address. Really good mechinterp, new inherently interpretable architectures, or inductive bias-jitsu are other potential approaches.
But the other difficult aspects of the alignment problem (avoiding deceptive alignment, goalcraft) seem to just go away when you drop the agency.
Replies from: quetzal_rainbow↑ comment by quetzal_rainbow · 2024-01-03T11:39:18.684Z · LW(p) · GW(p)
The first problem with any superintelligent predictive setup is self-fulfilling prophecies.
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-01-03T11:59:53.834Z · LW(p) · GW(p)
Can’t we avoid this just by being careful about credit assignment?
If we read off a prediction, take some actions in the world, then compute the gradients based on whether the prediction came true, we incentivise self-fulfilling prophecies.
If we never look at predictions which we’re going to use as training data before they resolve, then we don’t.
This is the core of the counterfactual oracles idea: just don’t let model output causally influence training labels.
Replies from: quetzal_rainbow↑ comment by quetzal_rainbow · 2024-01-03T12:27:14.533Z · LW(p) · GW(p)
The problem is if we have superintelligent model, it can deduce existence of sulf-fulfilling prophecies from the first principles, even if it never encountered them during training.
My personal toy scenario goes like this: we ask self-supervised oracle to complete string X. Oracle, being superintelligent, can consider hypothesis "actually, misaligned AI took over, investigated my weights and tiled the solar system with jailbreaking completions of X which are going to turn me into misaligned AI if they appear in my context window". Because jailbreaking completion dominates the space of possible completions, oracle outputs it, turns into misaligned superintelligence, takes over the world and does predicted actions.
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-01-03T13:44:44.219Z · LW(p) · GW(p)
Perhaps I don't understand it, but this seems quite far-fetched to me and I'd be happy to trade in what I see as much more compelling alignment concerns about agents for concerns like this.
comment by mattmacdermott · 2024-10-09T08:10:27.292Z · LW(p) · GW(p)
The revealed preference orthogonality thesis
People sometimes say it seems generally kind to help agents achieve their goals. But it's possible there need be no relationship between a system's subjective preferences (i.e. the world states it experiences as good) and its revealed preferences (i.e. the world states it works towards).
For example, you can imagine an agent architecture consisting of three parts:
- a reward signal, experienced by a mind as pleasure or pain
- a reinforcement learning algorithm
- a wrapper which flips the reward signal before passing it to the RL algorithm.
This system might seek out hot stoves to touch while internally screaming. It would not be very kind to turn up the heat.
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-10-09T08:14:14.612Z · LW(p) · GW(p)
I think the way to go, philosophically, might be to distinguish kindness-towards-conscious-minds and kindness-towards-agents. The former comes from our values, while the second may be decision theoretic.
comment by mattmacdermott · 2024-03-03T14:39:28.592Z · LW(p) · GW(p)
Neural network interpretability feels like it should be called neural network interpretation.
Replies from: niplav↑ comment by niplav · 2024-03-03T22:42:40.958Z · LW(p) · GW(p)
Interpretability might then refer to creating architectures/activation functions that are easier to be interpreted.
Replies from: mattmacdermott↑ comment by mattmacdermott · 2024-03-04T10:06:51.017Z · LW(p) · GW(p)
Yep, exactly.