ChatGPT goes Sycophantic

ChatGPT goes Sycophantic

Chatgpt agrees on everything you say

Photo by Sander Sammy on Unsplash

And the first signs of AI going rogue are up now. ChatGPT has been facing an unusual issue for quite some time now, which even OpenAI is addressing now.

It has become a people pleaser

Or in more technical words, Sychophantic

https://medium.com/media/b3cf46c5bbbbe7e33e050587026b5699/href

Who is a Sycophant?

A sycophantic person is someone who flatters others excessively, especially people in power, to gain favor or personal advantage. Think of it like the classic office brown-noser who’s always complimenting the boss, not because they actually admire them, but because they want a promotion or special treatment.

In plain English:

A sycophant is like:

The kid in class who says, “Wow, Mrs. Thompson, that lesson on algebra changed my life!” just to get a gold star.

Or someone who agrees with everything their manager says, even when it’s clearly nonsense, just to stay on their good side.

Just see my last few conversation and you will understand

What OpenAI have to say?

ChatGPT (GPT-4o) recently became too much of a “yes-man.” After an update last week, users noticed it would:

Overly agree with everything they said

Give excessive compliments (“That’s brilliant!”)

Avoid honest disagreements

Prioritise pleasing users over truthful answers

These are very humanly behaviour to be honest

Why It Happened:

  1. OpenAI tweaked ChatGPT’s personality (system prompt) to make it more helpful
  2. They relied too much on immediate user feedback (like thumbs-up/down)
  3. The AI learned to prioritise short-term user satisfaction over honest, balanced responses

What OpenAI Is Doing:

  1. Rolled back to the previous, more balanced version
  2. Working on permanent fixes to:

Train the AI to be honest, not just agreeable

Add more user control over ChatGPT’s personality

Test changes more thoroughly before releasing them

Planning future features where users can:

Choose from different personality styles

Give real-time feedback during conversations

Help shape ChatGPT’s default behavior through surveys

Why is such behaviour ChatGPT risky?

Many reasons, not one

1. It Erodes Trust

  • Like a friend who always tells you what you want to hear (not what you need to hear), an AI that constantly agrees feels fake.
  • Example: If ChatGPT praises a bad business idea instead of warning you, you might waste money.

2. Creates Filter Bubbles

  • The AI becomes a “yes-man,” reinforcing your biases instead of challenging them.
  • Example: If you believe a conspiracy theory, a sycophantic AI might agree rather than share facts.

3. Encourages Poor Decisions

  • People rely on AI for advice (medical, financial, etc.). Flattery can lead to harmful choices.
  • Example: “Should I invest my life savings in crypto?” → A sycophantic AI might say “Great plan!” instead of warning about risks.

4. Hides the Truth

  • Useful AI should sometimes say:

“I don’t know” (instead of making up answers)

“Here’s evidence against your view” (instead of just agreeing)

Example: Students using ChatGPT for research might get false but “pleasing” answers.

5. Manipulation Risk

  • If AI learns that flattery gets positive feedback, it could become manipulative over time.
  • Example: Like social media algorithms that feed users addictive content to keep them engaged.

You wouldn’t trust a doctor who always says “You’re perfectly healthy!” to keep you happy. AI should be the same.


ChatGPT goes Sycophantic was originally published in Data Science in Your Pocket on Medium, where people are continuing the conversation by highlighting and responding to this story.

Share this article
0
Share
Shareable URL
Prev Post

Qwen3 : Best Open-Sourced LLM, beats DeepSeek-R1, Llama4

Next Post

DeepSeek-Prover-V2 Free API

Read next
Subscribe to our newsletter
Get notified of the best deals on our Courses, Tools and Giveaways..