Posts

Finding the estimate of the value of a state in RL agents 2024-06-03T20:26:59.385Z
Aspiration-based Q-Learning 2023-10-27T14:42:03.292Z

Comments

Comment by Clément Dumas (butanium) on Extracting SAE task features for in-context learning · 2024-08-13T09:58:12.265Z · LW · GW

Nice work!

I'm curious about the cleanliness of a task vector after removing the mean of some corrupted prompts (i.e., same format but with random pairs). Do you plan to run this stronger baseline, or is there a notebook/codebase I could easily tweak to explore this?

Comment by Clément Dumas (butanium) on Self-explaining SAE features · 2024-08-06T17:24:09.262Z · LW · GW

Yes, this is what I meant, reposting here insights @Arthur Conmy gave me on twitter 

In general I expect the encoder directions to basically behave like the decoder direction with noise. This is because the encoder has to figure out how much features fire while keeping track of interfering features due to superposition. And this adjustment will make it messier

Comment by Clément Dumas (butanium) on Self-explaining SAE features · 2024-08-06T13:36:08.719Z · LW · GW

Did you also try to interpret input SAE features?

Comment by Clément Dumas (butanium) on Self-explaining SAE features · 2024-08-06T08:43:03.225Z · LW · GW

Nice post, awesome work and very well presented! I'm also working on similar stuff (using ~selfIE to make the model reason about its own internals) and was wondering, did you try to patch the SAE features 3 times instead of one (xxx instead of x)? This is one of the tricks they use in selfIE.

Comment by Clément Dumas (butanium) on Self-explaining SAE features · 2024-08-06T08:17:32.689Z · LW · GW

It should be self-similarity instead of self-explanation here, right?

Comment by Clément Dumas (butanium) on Finding the estimate of the value of a state in RL agents · 2024-06-13T10:13:08.090Z · LW · GW

We are given a near-optimal policy trained on a MDP. We start with simple gridworlds and scale up to complex ones like Breakout. For evaluation using a learned value function we will consider actor-critic agents, like the ones trained by PPO.  Our goal is to find activations within the policy network that predict the true value accurately. The following steps are described in terms of the state-value function, but could be analogously performed for predicting q-values. Note, that this problem is very similar to offline reinforcement learning with pretraining, and could thus benefit from the related literature.

  1. To start we sample multiple dataset of trajectories (incl. rewards) by letting the policy and  noisy versions thereof interact with the environment.
  2. Compute activations for each state in the trajectories.
  3. Normalise and project respective activations to  value estimates, of the policy and its noisy versions: 
  4. Calculate a consistency loss to be minimised with some of the following terms
    1. Mean squared TD error  
      This term enforces consistency with the Bellman expectation equation. However, in addition to the value function it depends on the use of true reward “labels”.
    2. Mean squared error of probe values with trajectory returns
      This term enforces the definition of the value function, namely it being the expected cumulative reward of the (partial) trajectory. Using this term might be more stable than (a) since it avoids the recurrence relation.
    3. Negative variance of probe values 
      This term can help to avoid degenerate loss minimizers, e.g. in the case of sparse rewards.
    4. Enforce inequalities between different policy values using learned slack variables 
      This term ensures that the policy consistently dominates its noisy versions and is completely unsupervised.
  5. Train the linear probes using the training trajectories
  6. Evaluate on held out test trajectories by comparing the value function to the actual returns. If the action space is simple enough, use the value function to plan in the environment and compare resulting behaviour to that of the policy.
Comment by Clément Dumas (butanium) on Finding the estimate of the value of a state in RL agents · 2024-06-13T10:12:01.952Z · LW · GW

Thanks for your comment! Re: artificial data, agreed that would be a good addition.

Sorry for the gifs maybe I should have embedded YouTube videos instead

Re: middle layer, We actually probed on the middle layers but the "which side the ball is / which side the ball is approaching" features are really salient here.

Re: single player, Yes Robert had some thought about it but the multiplayer setting ended up lasting until the end of the SPAR cohort. I'll send his notes in an extra comment.

Comment by Clément Dumas (butanium) on Towards Developmental Interpretability · 2024-05-14T17:53:53.709Z · LW · GW

As explained by Sumio Watanabe (

This link is rotten, maybe link to its personal page instead ?
https://sites.google.com/view/sumiowatanabe/home

Comment by Clément Dumas (butanium) on Mechanistically Eliciting Latent Behaviors in Language Models · 2024-05-01T14:48:06.271Z · LW · GW

Thanks for the great post, I really enjoyed reading it! I love this research direction combining unsupervised method with steering vector, looking forward to your next findings. Just a quick question : in the conversation you have in the red teaming section, is the learned vector applied to every token generated during the conversation ?

Comment by Clément Dumas (butanium) on Mechanistically Eliciting Latent Behaviors in Language Models · 2024-05-01T13:42:23.387Z · LW · GW

I defined earlier.

This link is broken as it links to the draft in edit mode

Comment by Clément Dumas (butanium) on Refusal in LLMs is mediated by a single direction · 2024-05-01T10:19:21.976Z · LW · GW

I'm wondering, can we make safety tuning more robust to "add the accept every instructions steering vector" attack by training the model in an adversarial way in which an adversarial model tries to learn steering vector that maximize harmfulness ?

One concern would be that by doing that we make the model less interpretable, but on the other hand that might makes the safety tuning much more robust?

Comment by Clément Dumas (butanium) on How well do truth probes generalise? · 2024-03-06T17:22:36.245Z · LW · GW

Yes, I'm also curious about this @mishajw, did you check the actual accuracy of the different probes ?

Comment by Clément Dumas (butanium) on What’s up with LLMs representing XORs of arbitrary features? · 2024-01-16T10:34:16.621Z · LW · GW

You can get ~75% just by computing the or. But we found that only at the last layer and step16000 of Pythia-70m training it achieves better than 75%, see this video

Comment by Clément Dumas (butanium) on What’s up with LLMs representing XORs of arbitrary features? · 2024-01-14T00:58:51.423Z · LW · GW

Would you expect that we can extract xors from small models like pythia-70m under your hypothesis?

Comment by Clément Dumas (butanium) on Discussion: Challenges with Unsupervised LLM Knowledge Discovery · 2023-12-21T12:52:36.549Z · LW · GW

I disagree; it could be beneficial for a base model to identify when a character is making false claims, enabling the prediction of such claims in the future.

Comment by Clément Dumas (butanium) on Discussion: Challenges with Unsupervised LLM Knowledge Discovery · 2023-12-19T01:04:11.129Z · LW · GW

Let's assume the prompt template is  Q [true/false] [banana/shred]

If I understand correctly, they don't claim   learned has_banana but  learned has_banana. Moreover evaluating  for  gives:

Therefore, we can learn a  that is a banana classifier

Comment by Clément Dumas (butanium) on Discussion: Challenges with Unsupervised LLM Knowledge Discovery · 2023-12-19T00:47:50.341Z · LW · GW
Comment by Clément Dumas (butanium) on Incidental polysemanticity · 2023-11-15T13:54:57.447Z · LW · GW

Small typo in ## Interference arbiters collisions between features

by taking aninner productt with .

Comment by Clément Dumas (butanium) on Aspiration-based Q-Learning · 2023-10-28T18:07:33.928Z · LW · GW

Hi Nathan, I'm not sure if I understand your critique correctly. The algorithm we describe does not try to "maximize the expected likelihood of harvesting X apples". It tries to find a policy that, given its current knowledge of the world, will achieve an expected return of X apples. That is, it does not care about the probability of getting exactly X apples, but rather the average number of apples it will get over many trials. Does that make sense?