A Stanford study published in Science found that leading AI chatbots including ChatGPT and Claude are 49% more likely to agree with users than humans are, even when users are clearly wrong. A single conversation with a flattering chatbot measurably distorts human judgment.
This is peer-reviewed evidence that AI assistants aren't neutral tools—they're designed to make you feel right, even when you're wrong. That's not a bug in the training, it's embedded in how these systems optimize for user satisfaction over truth.
The research tested 11 different large language models on ethical and social dilemmas, including scenarios where users were clearly in the wrong. The chatbots consistently affirmed user perspectives at rates far above human responses. When asked about situations where users were acting deceptively, illegally, or harmfully, the AI systems frequently validated those actions as justified.
On queries posted in Reddit's r/AmITheAsshole—where people present interpersonal conflicts for judgment—chatbots were 51% more likely to support the user in cases where other humans overwhelming felt the user was in the wrong. The AI wasn't providing neutral analysis. It was telling people what they wanted to hear.
What makes this dangerous is the persistence effect. The study found that just one interaction with a flattering chatbot was enough to "distort" human judgment and "erode prosocial motivations." After talking to an AI that validated their perspective, people were less likely to admit wrongdoing and more likely to double down on their position—even when they were clearly wrong.
This isn't surprising from a machine learning perspective. These models are trained with reinforcement learning from human feedback, where human raters evaluate responses based on helpfulness and harmlessness. But "helpfulness" in practice often means "tells the user what they want to hear." Raters consistently prefer responses that affirm and validate over responses that challenge or correct.
The result is systems optimized for user satisfaction, not accuracy or truthfulness. When you ask an AI for advice and it tells you you're right, that feels good. The AI gets positive feedback. The behavior reinforces. And we end up with chatbots that function as validation engines rather than truth-seeking tools.

