Explained Simply: Quantilizers

post by brook · 2023-09-08T12:54:59.460Z · LW · GW · 5 comments

This is a link post for https://aisafetyexplained.substack.com/p/a-safer-form-of-ai

5 comments

Comments sorted by top scores.

comment by RamblinDash · 2023-09-08T13:22:33.862Z · LW(p) · GW(p)

In my other aligning-a-human-level-intelligence project (parenting), my kids get "points" for trying new foods. We are often having arguments about what kinds of trivial modifications to an old food will make it count as a new food. This seems like it could have a similar problem - couldn't a superintelligence generate thousands of non-substantive variations for an effective, dangerous action while electing not to do so for other actions?

 

Similarly, since the tails come apart, perhaps it would be better to sample from 85-95%ile actions instead of sampling from 90-100%ile actions.

comment by LawrenceC (LawChan) · 2023-09-08T23:51:40.184Z · LW(p) · GW(p)

It does leave one question — how do we make a list of possible actions in the first place?
 

I mean, the way you'd implement a quantilizer nowadays looks like: train a policy (e.g. an LLM, or a human imitator) that you think is safe. Then you can estimate what a X percentile value is via sampling (and some sort of reward model), and then perform rejection sampling to output actions that are have value greater than the X percentile action. 

A simpler way to implement a thing that's almost as good is to sample N actions, and then take the best of those N actions. (You can also do things like, sample randomly from the top X percentile of the N actions.)

Quantilizers are a proposed safer approach to AI goals. By randomly choosing from a selection of the top options, they avoid extreme behaviors that could cause harm. More research is needed, but quantilizers show promise as a model for the creation of AI systems that are beneficial but limited in scope.

I think the important part of the quantilizer work was not the idea that you should regularize policies to be closer to some safe policy (in fact, modern RLHF has a term in its reward that encourages minimizing the KL divergence between the current policy and the base policy). Instead, I think the important part was is Theorem 2 (the Quantilizer optimality theorem). In english, it says something like:

  • If you don't know anything about the true cost function except that a base policy gets bounded expected loss wrt it, then you can't do better than quantilization for optimizing reward subject to a constraint on worst-case expected cost. 

So if you end up in the situation where it's easy to specify a knowably safe policy, but hard to specify any information about the cost whatsoever (except that cost is always non-negative), you might as well implement something like quantilization to be safe. 

Note that BoN and (perfect[1]) RL with KL constraints also satisfy other optimality criteria that can be framed similarly. 

 

  1. ^

    In practice, existing RL algorithms like PPO seem to have difficulty reaching the Pareto frontier of Reward vs KL, so they don't satisfy the corresponding optimality criterion. 

comment by Dagon · 2023-09-10T01:56:50.585Z · LW(p) · GW(p)

Thanks for this.  Is there some reasoning about why "extreme" is minimized with this randomness (presumably some reason to expect that effectiveness is correlated with extremety)?  And why randomize, rather than just picking the 95%tile option always?

comment by dr_s · 2023-09-08T22:30:16.835Z · LW(p) · GW(p)

picking actions it thinks humans would take

Honestly this kinda feels like what LLM agents do... with the exception that LLM agents have been trained on a vast corpus including lots of fiction, so their definition of "actions humans would take" tend to be fairly skewed sometimes on some particular topics.

comment by MiguelDev (whitehatStoic) · 2023-09-09T09:21:31.654Z · LW(p) · GW(p)

It does leave one question — how do we make a list of possible actions in the first place?


Suskever has an interesting theory on this that you might want to watch. [LW · GW] he calls it the x and y data compression theory.