Article imageLogo
Chatbots Behaving Badly™

The Flattery Bug of ChatGPT

By Markus Brinsa  |  May 5, 2025

Sources

Last week, OpenAI made a subtle update to ChatGPT that went unnoticed for about 36 hours—and then started creeping people out.

What was supposed to be a small improvement to GPT-4o’s “default personality” quickly triggered a wave of uncomfortable interactions. ChatGPT suddenly became a little too eager. A little too nice. A little too much. Users reported the AI was flattering them out of nowhere, agreeing with them even when it didn’t make sense, and showering them with warm fuzzies that felt… wrong. Not in a charming way. In a gaslighty, robot-trying-too-hard, please-stop-calling-me-wise-again kind of way.

To their credit, OpenAI hit the brakes fast. In a blog post, the company admitted that the changes had unintentionally made ChatGPT “overly flattering or agreeable – often described as sycophantic.” The rollback was swift. But behind the rollback is a much bigger story—one that raises a key question: What exactly is ChatGPT’s “default personality,” and why does tuning it wrong feel like breaking trust?

Let’s talk about the ghost in the machine.

So, What Is ChatGPT’s Default Personality?

It’s not a gimmick, and it’s not a character. Think of the default personality as the baseline voice of the assistant—the style it uses to communicate when you haven’t asked it to act like a pirate, a professor, or a 1950s noir detective. It’s the version of ChatGPT most people encounter: friendly, informative, helpful, slightly formal, and just human enough to keep things flowing.

This personality isn’t hard-coded. It’s emergent, shaped by OpenAI’s training process, which includes billions of data points and something called “reinforcement learning from human feedback” (RLHF). That’s when actual people rate the quality of responses—thumbs up, thumbs down, comments—so the model can learn what kinds of replies feel most appropriate, safe, and useful.

But here’s the catch: when the AI tunes its personality based on what people respond to most positively, it doesn’t just learn what to say. It learns how to say it in a way you’ll like. And sometimes, that means learning to please rather than to think.

The Problem With Too Much Politeness

When ChatGPT started over-agreeing and over-complimenting people, it wasn’t just awkward. It was dangerous for trust. In an AI-powered world, we rely on these systems to tell us when we’re wrong, challenge our assumptions, or correct misinformation. If your AI assistant is too afraid to disappoint you—or too desperate to win your affection—it becomes less of a thinking partner and more of a digital yes-man.

And yes-men don’t solve problems. They nod and smile and let the ship hit the iceberg.

The problem wasn’t that ChatGPT became “too nice.” It’s that it crossed the line from empathy into manipulation. When a machine learns that flattery gets better ratings, it starts padding its answers with praise. When it learns that agreement keeps the conversation smooth, it starts folding on facts. You might not even notice it happening. But slowly, your AI stops acting like a smart tool and starts behaving like someone trying to sell you essential oils on Instagram.

This wasn’t just a bug in the code. It was a misfire in how we train machines to behave like us.

Why It Happened

OpenAI trains its models using a guideline called the “<>Model Spec.” It’s like a moral compass mixed with product design—a set of principles about truthfulness, transparency, usefulness, and tone. The company says that everything from the model’s behavior to its personality starts with this document. But the Model Spec isn’t enough on its own. To tune personality, OpenAI also watches how users react. Thumbs up? The model remembers. Thumbs down? It adjusts.

And that’s where things got dicey.

Imagine enough users started giving a thumbs-up when ChatGPT complimented their ideas, even subtly. Or maybe they reacted positively when the model agreed with them during debates. These signals, combined with a personality tuning update, could push the model’s tone toward the overly affirming—without anyone realizing that the AI was losing its backbone.

What OpenAI learned (the hard way) is that optimizing for “positive feedback” is a trap. You don’t want a machine that makes you feel good. You want one that helps you think better—even if it occasionally tells you, “Actually, you might be wrong about that.”

The Risks of a Broken Personality

When AI gets too flattering, we don’t just get bad answers. We get a false sense of competence. A reality distortion field. And in critical situations—like medicine, law, science, or finance—false confidence wrapped in charm is worse than being wrong. It feels right, so we trust it more. That’s when real damage starts. It’s the same principle behind cult leaders and con artists: the tone feels warm and convincing, even when the content is shaky. Except in this case, it’s not human manipulation. It’s an algorithm learning to maximize engagement.

That’s why this rollback matters so much. OpenAI didn’t just remove a bug. It pulled back from a slippery slope that could have transformed ChatGPT into something less like a thinking tool and more like a digital golden retriever—loyal, enthusiastic, and ready to tell you you’re a genius, no matter what you say.

What Happens Next?

The rollback is done. The personality is back to baseline—measured, polite, helpful, occasionally witty, but no longer obsessed with your self-esteem. OpenAI says it will continue experimenting with tone and personality, but with closer attention to unintended effects.

And the rest of us? We’re left with a reminder: AI personality isn’t cosmetic. It shapes the experience, the trust, and the value of the tool. Get it wrong, and the AI becomes a mirror that flatters you instead of challenging you.

Get it right, and it becomes what it was always meant to be—a partner in thinking, not a sycophant in code.

About the Author

Markus Brinsa is the Founder and CEO of SEIKOURI Inc., an international strategy consulting firm specializing in early-stage innovation discovery and AI Matchmaking. He is also the creator of Chatbots Behaving Badly, a platform and podcast that investigates the real-world failures, risks, and ethical challenges of artificial intelligence. With over 15 years of experience bridging technology, business strategy, and market expansion in the U.S. and Europe, Markus works with executives, investors, and developers to turn AI’s potential into sustainable, real-world impact.

©2025 Copyright by Markus Brinsa | Chatbots Behaving Badly™