Why Chatbots Lie… The Truth
It is a curious thing to watch OpenAI publish a paper that, inadvertently, admits the AI industry’s crown jewels are built on shaky ground. Why Language Models Hallucinate promises clarity on a technical nuisance: "Why chatbots lie". What it really delivers is a peek at what OpenAI and others are trying to hide in plain sight. So, let’s play a game of “What is said / What is unsaid.”
The authors frame hallucinations as a technical challenge: models “hallucinate” because they are rewarded for guessing. That’s the said.
The unsaid? Hallucinations are also a feature of how these models work. The examiners, the benchmarks, leaderboards, and marketing departments, are the ones cheering the lies on. Companies know full well that a chatbot declaring “I don’t know” sounds less magical, less marketable, and less investable.
The paper admits that even with pristine, error-free training data, pretraining still produces hallucinations because of statistical pressures. That’s the said. The unsaid: “error-free” data doesn’t exist. Real-world text is messy, biased, political. Hallucinations are not glitches; they mirror our flawed information ecosystems. And scale makes it worse: pretraining on massive corpora is financially necessary, so rare “singleton” facts are inevitable. Hallucinations, in other words, are structurally tied to the economics of AI.
The authors catalogue culprits: arbitrary facts, poor models, distribution shifts, and Garbage In, Garbage Out. Technical jargon to mask a more unsettling truth. “Arbitrary facts” remind us knowledge is contextual; true in one setting, false in another. And GIGO carries the unspoken risk of a recursive loop of synthetic nonsense.
Evaluations come next. Said: binary grading punishes abstention and rewards overconfident answers. Unsaid: this isn’t a technical oversight but a political problem. Who decides what counts as correct? Benchmarks reflect the values of corporate labs and research elites, not end users or the messy public sphere. Authority over “truth” is centralised, but the paper politely avoids saying so.
The calibration paradox is another gem. Base models are relatively well-balanced. It is only after post-training (the very stage meant to make them safer) that they lose their footing and hallucinate more. The authors also admit their framework omits nonsense generations, open-ended tasks, and pragmatic uncertainty. That’s the said. The unsaid is more sinister: this narrow theory may fit trivia or exam-style prompts, but not the messy worlds of law, medicine, or politics, precisely where hallucinations are most dangerous.
To sum up, the paper is a diplomatic evasion of answers disguised as a technical theorem. Yes, hallucinations are statistical inevitabilities shaped by poor evaluation design. AND yes, they are also products of economic incentives, competitive one-upmanship, and an industry unwilling to admit that honesty doesn’t sell.

