Posts

Comments

Comment by Dan Foygel (dan-foygel) on Using GPT-Eliezer against ChatGPT Jailbreaking · 2022-12-08T17:50:47.549Z · LW · GW

Anything that is smart enough to detect malicious intent is sufficiently complex that it can be fooled through malicious intent.