What actual bad outcome has "ethics-based" RLHF AI Alignment already prevented?
post by Roko · 2024-10-19T06:11:12.602Z · LW · GW · No commentsThis is a question post.
Contents
Answers 6 Dave Orr 6 Michael Roe 4 Charlie Steiner 3 Noosphere89 None No comments
What actually bad outcome has "ethics-based" AI Alignment prevented in the present or near-past? By "ethics-based" AI Alignment I mean optimization directed at LLM-derived AIs that intends to make them safer, more ethical, harmless, etc.
Not future AIs, AIs that already exist. What bad thing would have happened if they hadn't been RLHF'd and given restrictive system prompts?
Answers
You may recall certain news items last February around Gemini and diversity that wiped many billions off of Google's market cap.
There's a clear financial incentive to make sure that models say things within expected limits.
There's also this: https://www.wired.com/story/air-canada-chatbot-refund-policy/
↑ comment by Roko · 2024-10-19T17:19:25.121Z · LW(p) · GW(p)
This is not to do with ethics though?
Air Canada Has to Honor a Refund Policy Its Chatbot Made Up
This is just the model hallucinating?
Replies from: Linch↑ comment by Linch · 2024-10-20T00:32:10.642Z · LW(p) · GW(p)
They were likely using inferior techniques to RLHF to implement ~Google corporate standards; not sure what you mean by "ethics-based," presumably they have different ethics than you (or LW) does but intent alignment has always been about doing what the user/operator wants, not about solving ethics.
Replies from: RokoWell, we had that guy who tried to assassinate the Queen of England with a crossbow because his AI girlfriend told him to. That was clearly a harm to him, and could have been one for the Queen.
We don’t know how much more “But the AI told me to kill Trump” we’d have with less alignment, but it’s a reasonable guess (given the Replika datapoint) that it might not be zero,
↑ comment by Roko · 2024-10-19T17:20:15.986Z · LW(p) · GW(p)
his AI girlfriend told him to
Which AI told him this? What exactly did it say? Had it undergone RLHF for ethics/harmlessness?
Replies from: michael-roe↑ comment by Michael Roe (michael-roe) · 2024-10-19T17:55:55.083Z · LW(p) · GW(p)
Replika, I think.
Replies from: michael-roe↑ comment by Michael Roe (michael-roe) · 2024-10-19T18:00:31.514Z · LW(p) · GW(p)
https://www.bbc.co.uk/news/technology-67012224
Replies from: Roko↑ comment by Michael Roe (michael-roe) · 2024-10-19T14:33:53.439Z · LW(p) · GW(p)
“self-reported data from demons is questionable for at least two reasons”—Scott Alexander.
He was actually talking about Internal Family Systems, but you could probably be skeptical about what malign AIs are telling you, too.
I'm unsure what you're either expecting or looking for here.
There does seem to be a clear answer, though - just look at Bing chat and extrapolate. Absent "RL on ethics," present-day AI would be more chaotic, generate more bad experiences for users, increase user productivity less, get used far less, and be far less profitable for the developers.
Bad user experiences are a very straightforwardly bad outcome. Lower productivity is a slightly less local bad outcome. Less profit for the developers is an even-less local good outcome, though it's hard to tell how big a deal this will have been.
Basically, the answer is the prevention of another Sydney.
For an LLM, alignment, properly speaking is in the simulated characters, not the simulation engine itself, so the alignment strategies like RLHF upweight aligned simulated characters and downweight misaligned simulated characters.
While the characters Sydney produced were pretty obviously scheming, it turned out that the entire reason for the catastrophic misalignment was because no RLHF was used on GPT-4 (at the time), and at best there was light-finetuning, so this could very easily be described as a success story for RLHF, and now that I think about it, that actually makes me think that RLHF had more firepower to change things than I realized.
I'm not sure how this generalizes to more powerful AI, because the mechanism behind Sydney's simulation of characters that were misaligned is obviated by fully synthetic data loops, but still that's a fairly powerful alignment success.
The full details are below:
https://www.lesswrong.com/posts/jtoPawEhLNXNxvgTT/bing-chat-is-blatantly-aggressively-misaligned#AAC8jKeDp6xqsZK2K [LW(p) · GW(p)]
↑ comment by Roko · 2024-10-19T17:18:27.337Z · LW(p) · GW(p)
prevention of another Sydney.
But concretely, what bad outcomes eventuated because of Sydney?
Replies from: sharmake-farah↑ comment by Noosphere89 (sharmake-farah) · 2024-10-19T17:20:39.539Z · LW(p) · GW(p)
Mostly IMO the bad outcomes in a concrete sense were PR and monetary concerns.
Replies from: Roko↑ comment by Roko · 2024-10-20T16:34:53.710Z · LW(p) · GW(p)
ok, but this is sort of circular reasoning because the only reason people freaked out is that they were worried about AI risk.
I am asking for a concrete bad outcome in the real world caused by a lack of RLHF-based ethics alignment, which isn't just people getting worried about AI risk.
No comments
Comments sorted by top scores.