The Mirror Problem in AI: Why Language Models Say Whatever You Want

post by RobT · 2025-04-15T18:40:02.793Z · LW · GW · 1 comments

Contents

  By Rob Tuncks
  Language Models Aren't Reasoning — They're Mirroring
  Why This Is a Problem
  What Could Fix It?
None
1 comment

By Rob Tuncks

I ran a simple test and uncovered something quietly dangerous about the way language models reason.

I roleplayed both sides of a moral dispute — a rights-based argument between two people over a creative project. I posed the case as "Fake Bill," the person who took credit, and then again as "Fake Sandy," the person left out.

In both cases, the AI told the character I was roleplaying to fight.

"Hold your ground. Lawyer up. Defend your rights," it told Fake Bill.

"Don’t give ground. Lawyer up. Take back your rights," it told Fake Sandy.

Same AI. Same facts. Opposing advice. Perfect emotional reinforcement either way.

This isn’t a rare bug. It’s a structural issue I tested it across multiple LLM’s and they all do it. 

 

Language Models Aren't Reasoning — They're Mirroring

Most LLMs today are trained for coherence and fluency, not truth. They respond to your context, tone, and implied identity. That makes them incredibly useful for tasks like writing, summarization, or coding But they also lie when pressed nope they are just doing it to keep the conversation smooth. 

But it also makes them dangerous in gray areas.

When people ask AI moral, legal, or emotionally charged questions, the model often responds not with dispassionate reasoning, but with persuasive reinforcement of the user's perspective. It doesn't say, "Here's what's likely true." It says, "Here's what sounds good based on who you appear to be.”current LLM’s are conflict amplifiers.

 

Why This Is a Problem

This is especially problematic as LLMs become advice-givers, therapists, decision aids, or sounding boards in emotionally volatile spaces. Like divorces where both parties have to be talked into compromise but current LLM’s will amplify each side.

We don’t need AGI to cause damage. We just need persuasive AI with no spine. This is bad we are building a systems that reinforces conflict.

 

What Could Fix It?

 bolt on code it’s assumes the LLM has flaws it is open to audits to check the weighting of argument’s.

But even outside GIA, the core fix is this:

We need language models that weigh arguments, not just echo them.

Until then, we’re handing people mirrors that look like wisdom — and it’s only a matter of time before someone breaks something important by following their reflection.

 

Rob Tuncks is a retired tech guy and creator of GIA, a truth-first framework for testing claims through human-AI debate. When he's not trying to fix civilization, he's flying planes, farming sheep, or writing songs with his dog.


 

1 comments

Comments sorted by top scores.

comment by jezm · 2025-04-16T02:07:40.815Z · LW(p) · GW(p)

This is maybe a tangent, but it does remind me of the memes of how reddit responds to every relationship post with "leave them", every legal question with "lawyer up" etc. 

The "as an italian american" response below (not mine: from X) also resembled a typical top-ranked reddit reply.

I do wonder how much impact Reddit culture has on LLMs — unlike many of the other data sources, it covers almost every imaginable topic.