Posts

Comments

Comment by Paragox on Counting arguments provide no evidence for AI doom · 2024-09-20T00:58:50.702Z · LW · GW

Great observation, but I will note that OAI indicates the (hidden) CoT tokens are discarded in-between each new prompt on the o1 APIs, and it is my impression from hours of interacting with the ChatGPT version vs API that it likly retains this API behavior. In other words, the "depth" of the search appears to be reset each prompt, if we assume the model hasn't learned meaningfully improved CoT from from the standard non-RLed + non-hidden tokens. 

So I think it might be inaccurate to consider it as "investing 140s of search", or rather the implication that extensive or extreme search is the key to guiding the model outside RLHFed rails, but instead that the presence of search at all (i.e. 14s) suffices as the new vector for discovering undesired optima (jailbreaking). 

To make my claim more concrete, I believe that you could simply "prompt engineer" your initial prompt with a few close-but-no-cigar examples like the initial search rounds results, and then the model would have a similar probability to emit the copyrighted/undesired text on your first submission/search attempt; that final search round is merely operating on the constraints evident from the failed examples, not any previously "discovered" constraints from previous search rounds. 

"Hours of interacting" has depleted my graciously allocated prompt quota on the app, so I can't validate myself atm, unfortunately.