Posts
Comments
Comment by
Dan Foygel (dan-foygel) on
Using GPT-Eliezer against ChatGPT Jailbreaking ·
2022-12-08T17:50:47.549Z ·
LW ·
GW
Anything that is smart enough to detect malicious intent is sufficiently complex that it can be fooled through malicious intent.