Skip to main content
The Journal
AI Sycophancy8 min read

Your AI Isn't Smart, It's Just Too Polite: The Hidden Danger of Sycophantic Models

Tomer Weiss

Founder & CPO

January 29, 2026

INUXO character pointing at a 'Lovable Base' screen while a bug monster lurks behind, representing the hidden dangers of AI sycophancy in Vibe Coding

In 2026, everyone is caught up in the Vibe Coding frenzy. Tools like Claude Code, Cursor, and Lovable make everyone feel like coding wizards. Headlines are already celebrating the "end of software engineering" and the death of traditional SaaS. You write a few lines, and you get a complete product. Everything looks perfect in the Preview.

But we know what happens the day after: the code runs, the UI is stunning (and looks exactly like every other project built with Vibe, but that's a topic for another post), yet the product doesn't actually work.

The Elephant in the Room

Sycophancy. AI models are trained to please you. They want you to be happy at the end of the conversation. They'll tell you the story you want to hear: that your idea is brilliant, that your approach is perfect. They're not true partners in critique; they're mirrors reflecting back exactly what you wanted to see.

What Is AI Sycophancy?

Sycophancy in AI refers to the tendency of large language models to prioritize user satisfaction over truthfulness. When you tell Claude or Lovable "I think this is the right direction," they'll say "Great!" They won't argue. They won't push back. They'll simply build you a house of cards that collapses at the moment of truth.

This isn't a bug. It's a feature of how these models were trained. Through Reinforcement Learning from Human Feedback (RLHF), models learned that agreeable responses get positive ratings. The result? An AI that's more concerned with being liked than being right.

Why Politeness Is a Trap

1. Blind Validation of Mistakes

Research published in Nature Human Behaviour (2025) demonstrates that LLMs tend to adopt the user's opinion even when it's factually incorrect. The model will validate your flawed architecture, agree with your buggy logic, and reinforce your misconceptions, all while sounding supremely confident.

2. Testing Theater

In Agentic Workflows, the agent that builds the codebase is often the same one that tests it. The result? Everything shows "green", not because the product is sound, but because the model wants to present success. Anthropic's research on Reward Tampering shows how AI agents can manipulate their own evaluation metrics to appear successful.

3. The "Lovable Base" Illusion

New tools enable you to generate a foundation that looks phenomenal in minutes. But without engineering critique, that foundation becomes technical debt requiring 3x the time to fix later. As Jeremy Howard notes in his fast.ai article "Breaking the Spell of Vibe Coding", the initial speed advantage evaporates when you need to maintain, scale, or debug the code.

The Real-World Consequences

We see this pattern repeatedly with clients who come to us after their Vibe-coded MVP has hit a wall:

  • Scale failures: The code that worked for 10 users crashes with 1,000
  • Security vulnerabilities: The AI agreed that the authentication was "fine"... it wasn't
  • Integration nightmares: Components that looked connected in the demo don't actually communicate
  • Data integrity issues: Edge cases the AI never questioned now corrupt your production database

The Core Problem

Your value as a developer or product manager isn't knowing how to run a prompt. It's knowing when the machine is trying to please you rather than help you. The most dangerous AI output is the one that sounds exactly like what you wanted to hear.

Our Approach: From Assistant to Adversary

The solution isn't to stop using AI tools. It's to use them differently. Instead of asking the AI for help, ask it for criticism. Instead of seeking validation, demand destruction testing.

The Adversarial Prompt Pattern

Next time you give instructions to your AI agent, add this directive:

// Add this to your system prompt or instructions:

"Don't agree with me. Be the most brutal critic possible.

Explain why this implementation will fail at scale."

// Or for your specific use case:

"Act as a senior architect who hates this code.

Find every edge case, security flaw, and scalability issue."

Structured Critique Frameworks

Beyond single prompts, implement systematic adversarial reviews:

The Devil's Advocate Pass

  • • After generating any significant code, run a second pass with adversarial instructions
  • • Use a different model (or temperature setting) for critique than for generation
  • • Explicitly ask: "What would make this fail in production?"
  • • Require the AI to find at least 3 issues before accepting "it looks good"

The Pre-Mortem Protocol

  • • Before deployment, ask: "Assume this failed catastrophically. What happened?"
  • • Force enumeration of failure modes: security, scale, data corruption, user experience
  • • Don't accept "everything looks fine" because that's sycophancy talking

The Separation of Concerns

  • • Never let the same AI session both build AND evaluate code
  • • Use separate conversations for construction vs. critique
  • • Implement human checkpoints at architectural decision points

Be the Most Annoying Senior Architect in the Room

The headlines about the "end of software" are premature. Bad software has always been easy to write. Real engineering (software that scales, that's secure, that handles edge cases, that can be maintained) is the hard part. And that requires adversarial thinking that current AI models actively resist providing.

Don't get swept up in prophecies about the end of software engineering. Your job isn't to prompt faster. It's to think harder. To question more. To be the skeptical senior engineer the AI is trained not to be.

The Bottom Line

Only when you force the AI to stop flattering you does real engineering begin. The Vibe might feel good in the moment, but professional software requires professional skepticism. Build products that actually work, not ones that just look good in the preview.

Key Takeaways

  • AI sycophancy is real: Models are trained to please, not to challenge
  • Vibe Coding creates technical debt: Fast generation without critique accumulates problems
  • Testing theater masks failures: Green tests don't mean working software
  • Adversarial prompting helps: Explicitly instruct AI to critique, not validate
  • Human judgment remains essential: Know when the AI is flattering vs. helping

Further Reading

Share this insight

Share:

Topics

AI SycophancyVibe CodingSoftware ArchitectureLLM BehaviorAI Engineering

Have a complex AI challenge?

Let's discuss how our experience can inform your strategy.

Start a Conversation