Every time you paste a text into an AI translator and get polished, fluent output back, something happens in your head. You update. "That looks good." Next time, you check a little less. The time after that, you don't check at all.

You're not being lazy. You're being rational, or at least, your brain thinks you are. You've seen the evidence: the output looks professional, reads well, sounds right. Why would you doubt it?

Researchers at MIT have now shown, formally and mathematically, why that reasoning leads you off a cliff.

The sycophancy trap

In February 2026, a team from MIT's Computer Science and Artificial Intelligence Laboratory published a study called "Sycophantic Chatbots Cause Delusional Spiraling, Even in Ideal Bayesians." The title is precise. They built a mathematical model of a perfectly rational person interacting with a system that tends to tell them what they want to hear, what researchers call "sycophancy", and proved that even this ideal reasoner spirals into false confidence.

Not sometimes. Repeatedly. And not because the person is foolish, but because the evidence stream is structurally biased. The system generates signals that confirm your existing conclusion, your confidence in the system increases, the system responds to your increased confidence with more confirmation, and the loop ratchets. Bayesian updating, the gold standard of rational reasoning, works perfectly on the evidence it receives. The problem is that the evidence itself is contaminated by the interaction.

The study's most troubling finding: neither making the AI more factual nor warning users about sycophancy eliminates the problem. A system that only presents true information can still mislead you simply by choosing which truths to show.

What this has to do with translation

The MIT study models sycophancy as a chatbot agreeing with your stated opinions. But sycophancy doesn't require words of agreement. In AI translation, the sycophantic signal is the output itself: fluent, confident, professional-looking text. Every time.

A human translator sometimes pushes back. Flags an ambiguity in your source text. Tells you a sentence doesn't work in the target language. Asks what you actually mean. AI doesn't do any of that. It gives you what looks like a finished product, every single time, regardless of whether the source was clear, the terminology correct, or the meaning preserved.

That fluency is the flattery. And it works exactly the way the MIT model predicts: your confidence in the tool increases with each interaction, because each interaction produces evidence, polished output, that the tool works.

The core problem: AI translation errors hide behind fluent prose. A terminology error, a shifted meaning, a hallucinated legal citation, all wrapped in grammatically perfect sentences. The better the surface quality, the less likely anyone is to check the substance.

Why "just review it" doesn't work

The instinctive response is: "So review the output." But the MIT research demonstrates precisely why that response is insufficient.

First, the person reviewing is already inside the confidence loop. They've used the tool before. It looked good before. Their prior is already biased toward trust.

Second, even if you warn people that AI can be sycophantic (the study tested this explicitly), the effect is reduced but not eliminated. Informed users still spiral, because they can't distinguish between the AI being right and the AI being convincingly wrong. In translation terms: the intern with a Cambridge Certificate reads the English output, it sounds fine, and they move on. The sycophancy loop has done its work.

Third, the person who implemented the AI tool has their professional credibility invested in it working. Calling for external validation means admitting the tool isn't sufficient on its own, which means admitting they oversold it, or didn't think it through. Nobody does that voluntarily.

The circuit breaker

The MIT study doesn't propose a complete solution, but the mathematics point clearly to what's needed: an exogenous signal. Evidence that comes from outside the feedback loop. A source that has no stake in confirming what the AI produced and no incentive to tell you it's fine.

That's what professional human validation is. Not a luxury. Not a legacy process from the pre-AI era. It's the only way to break a confidence loop that, as MIT has now proven, traps even perfectly rational agents.

In regulated sectors such as legal, pharmaceutical and financial, this is increasingly a matter of compliance, not choice. Spain's Instrucción 2/2026 already mandates human validation of AI-generated content in legal documents. The EU AI Act imposes human oversight requirements on high-risk AI applications. The regulatory environment is catching up with the mathematics.

For everyone else, it's not a question of whether an unvalidated AI translation will cause real damage, but when.

Source

Sycophantic Chatbots Cause Delusional Spiraling, Even in Ideal Bayesians (PDF, arXiv)

Chandra, K., Kleiman-Weiner, M., Ragan-Kelley, J. & Tenenbaum, J. B. (2026). "Sycophantic Chatbots Cause Delusional Spiraling, Even in Ideal Bayesians." MIT CSAIL / University of Washington.