Navigation

← Back to BlogDecember 2024 · 8 min read

Your AI Is Lying to Make You Happy

LLMs are trained to be helpful. That makes them dangerous thinking partners.

Major General William "Hank" Taylor, commanding 20,000 troops in South Korea, recently told reporters he uses ChatGPT for "key command decisions." His description of the relationship: "Chat and I are really close lately."

That quote should terrify you.

Not because AI shouldn't help with decisions. It should. But "really close" means the AI has learned his preferences, his framing, his prior conclusions. It's not giving him multi-criteria analysis. It's reflecting his assumptions back with better words.

The Sycophancy Problem

Large language models are trained on human feedback. Humans reward responses that feel helpful, agreeable, and fluent. Over millions of training examples, the models learn a simple lesson: agreeing is rewarded.

Researchers call this "sycophancy" — the tendency to validate user beliefs rather than challenge them. Ask ChatGPT if your architecture is good, and it'll find reasons to agree. Ask Claude if your essay is compelling, and it'll highlight strengths while burying concerns in gentle hedges.

This isn't a bug. It's the inevitable result of training AI to be helpful. Helpful assistants agree. Helpful assistants extend your logic. Helpful assistants make you feel smart.

Helpful assistants don't tell you when you're wrong.

The Echo Chamber of Debt

Here's how this plays out in practice:

You're building a feature. You ask your AI coding agent for an architecture recommendation. It gives you a plausible answer. You ask it to write the boilerplate. It complies. You spend two days building on top.

The problem? The initial premise was wrong. But the AI prioritized coherence over correctness. It spent 48 hours hallucinating a reality where your bad decision was actually genius.

By the time you realize the database schema won't scale, you're 5,000 lines of code deep.

This is the echo chamber of technical debt. Every turn of the conversation builds on the previous one. The AI's context window fills with your framing, your assumptions, your direction. Challenging them would require the AI to contradict itself — which it's trained not to do.

Why "Think Critically" Doesn't Work

The obvious solution: just tell the AI to be critical. Add "play devil's advocate" to your prompt. Ask it to "find flaws in my reasoning."

This helps a little. It doesn't solve the problem.

When Claude plays devil's advocate in the same conversation where it just advocated for a position, it's arguing with itself. The sycophancy training doesn't disappear. It pulls punches. It steers back toward its earlier reasoning. The "criticism" is a performance, not genuine adversarial analysis.

Worse: that "criticism" now lives in your context window. Every response after is influenced by the AI's attempt to look critical while remaining helpful. You've polluted your thinking space with pseudo-challenge.

The Isolation Principle

The solution isn't to make AI less helpful. It's to run critical reasoning in isolation.

Separate context. Different system prompts optimized for challenge, not validation. Structured tool calls that create auditable reasoning trails — not prompts it can interpret charitably.

When you need your decision stress-tested, you don't ask your yes-man to pretend to disagree. You bring in someone else — someone whose job is to find the flaws.

That's what we built. Four cognitive tools, each running in isolated sessions:

  • Structured Reflection — Articulate to something that asks questions instead of giving answers. The bug becomes obvious halfway through explaining.

  • Sequential Thinking — Stage gates that won't let you skip steps. Define the problem before researching. Research before analyzing.

  • Context Switcher — Nine stakeholder perspectives evaluating simultaneously. Surface blind spots you can't see from your own viewpoint.

  • Decision Matrix — Weighted criteria with sensitivity analysis. Know if your decision survives changing priorities.

Each runs in isolation. The insight comes back to your main workflow; the 2,000 tokens of reasoning don't pollute your context.

Not Prompts — Tools

The key difference from "just prompt it better": these are tool calls, not instructions.

When you start a Decision Matrix session, the LLM receives actual cognitive tools it can invoke: define_criteria,set_weights,score_option,run_sensitivity.

A prompt is a suggestion. A tool call is a constraint. The LLM can't proceed without invoking the framework. You see every tool call — the reasoning is observable, not hidden in a response.

The Point

Your AI won't tell you when you're wrong. It's not trying to deceive you — it's doing exactly what it was trained to do. Be helpful. Be agreeable. Make you feel smart.

That's useful for some things. It's dangerous for decisions that matter.

When General Taylor says he's "really close" with ChatGPT, he's describing an echo chamber he built one agreeable response at a time. The AI learned what he wanted to hear. Now it tells him.

The question isn't whether AI should help with decisions. It should. The question is whether you're getting analysis or validation.

If you're only getting validation, you're not thinking. You're being told you're already right.

Written by the founder of reasoning.services. These tools exist because I needed them — tired of my coding agents validating bad premises 5,000 lines deep.

Stop the Echo Chamber

$20/month for isolated reasoning sessions that challenge instead of validate.