The Journaly
Fact-Powered Stories · Est. 2026
5 min read
Artificial Intelligence

The AI That Lies With Confidence

Language models are smarter than ever — yet they still fabricate facts, citations, and entire realities without blinking.

March 24, 2026 · 3 hours ago · 5 min read

The AI That Lies With Confidence

In 2023, the engineers and researchers building the world's most powerful artificial intelligence systems made a bold promise: hallucinations — the industry's oddly whimsical term for when a chatbot simply invents facts — would soon be a solved problem. Two years later, the best AI models on the market still fabricate information between 3 and 18 percent of the time [7]. Worse, a disturbing finding from MIT researchers revealed that when these models hallucinate, they tend to do so using more confident language than when they're telling the truth [6]. The problem hasn't been solved. In many ways, it has deepened.

In 2023, the engineers and researchers building the world's most powerful artificial intelligence systems made a bold promise: hallucinations — the industry's oddly whimsical term for when a chatbot simply invents facts — would soon be a solved problem. Two years later, the best AI models on the market still fabricate information between 3 and 18 percent of the time [7]. Worse, a disturbing finding from MIT researchers revealed that when these models hallucinate, they tend to do so using more confident language than when they're telling the truth [6]. The problem hasn't been solved. In many ways, it has deepened.

The Confidence of a Liar

There is something uniquely unsettling about a machine that doesn't know what it doesn't know. When a large language model (LLM) hallucinates, it doesn't stutter, hedge, or trail off into ellipses. It delivers fabricated case law, invented scientific citations, and nonexistent historical events with the smooth authority of an expert witness. This is not a bug in the traditional sense — a misplaced semicolon or a corrupted file. It is something far more philosophically strange: a system optimized to sound correct, even when it isn't.

The term "hallucination" in the context of AI refers to outputs that are seemingly plausible and logical yet factually incorrect or unfaithful to the real world [9]. The label is almost too poetic. Human hallucinations are involuntary, sensory, often frightening. AI hallucinations are fluent, confident, and frequently undetectable without independent verification.

The New York Times reported in May 2025 that despite years of promises from developers, AI hallucinations are not only persisting — they are, by some measures, getting worse as models grow more powerful [5]. This is the central paradox of modern AI development: scale, which has delivered extraordinary leaps in capability, appears to introduce new and more sophisticated forms of confabulation. Researchers and developers told the Times in 2023 that the problem would be solved. It was not [5].

What makes this particularly alarming is the MIT finding, reported in early 2025, that models deploy more confident language precisely when they are hallucinating [6]. The system is not simply guessing — it is, in a structural sense, trained to fake answers it doesn't know [via Science.org reporting]. The reinforcement learning processes used to make these models more helpful and fluent may inadvertently reward confident-sounding responses, regardless of their accuracy. The model learns that hesitation is penalized. Certainty, even false certainty, is rewarded.

For everyday users, this creates a treacherous information environment. A lawyer who asked ChatGPT for case citations received a list of entirely fabricated rulings — a now-famous incident that resulted in court sanctions. A student submitting AI-assisted research may be citing papers that do not exist. The problem is not hypothetical. It is already reshaping how professionals, academics, and institutions must interact with these tools.

AI hallucinations — why language models still make things up - Why the Architecture Itself Is the Problem
Why the Architecture Itself Is the Problem — AI Generated
"When AI models hallucinate, they don't hesitate — they deliver fabricated facts with the smooth authority of an expert witness."

Why the Architecture Itself Is the Problem

AI hallucinations — why language models still make things up - The Industry's Uncomfortable Trade-Off
The Industry's Uncomfortable Trade-Off

To understand why hallucinations persist, you have to understand what a large language model actually is — and, crucially, what it is not. An LLM is not a database. It does not retrieve stored facts the way a search engine indexes pages. Instead, it is a statistical prediction engine, trained on vast quantities of human-generated text, that learns to predict which word, phrase, or sentence is most likely to follow a given input. It is, at its core, an extraordinarily sophisticated autocomplete system.

This architecture is the root of the hallucination problem. OpenAI's own research into why language models hallucinate points to a fundamental mismatch: the model's training process rewards fluency and coherence, not factual accuracy [3]. When a model encounters a question at the edge of its training data — something obscure, recent, or highly specialized — it doesn't have a mechanism to say "I don't know." It has a mechanism to generate a plausible-sounding continuation. And plausible is not the same as true.

Research published in ScienceDirect describes hallucination as emerging from the tension between the model's learned representations and the actual structure of reality [9]. The model has absorbed patterns of language so deeply that it can construct sentences that feel authoritative in any domain — medicine, law, history, science — without having reliable grounding in the facts of those domains. It has style without substance, in the worst cases.

The problem compounds as models grow larger. A 2025 Reddit thread discussing a research finding captured the paradox bluntly: as models scale up, they hallucinate more frequently, not less [8]. Larger models have more parameters, more nuanced representations, and a greater capacity to generate convincing prose in specialized fields. But that very sophistication makes their errors harder to detect. A small model might produce obviously garbled nonsense. A frontier model produces a beautifully written paragraph about a clinical trial that never happened.

Lakera's comprehensive guide to LLM hallucinations identifies several contributing factors beyond architecture alone: gaps and biases in training data, the model's inability to distinguish between what it "knows" confidently and what it is inferring, and the absence of any real-time grounding in verified information [4]. These are not minor implementation details. They are structural features of how these systems are built — which is precisely why the fix is so elusive.

"The model learns that hesitation is penalized. Certainty, even false certainty, is rewarded."

The Industry's Uncomfortable Trade-Off

Here is the uncomfortable truth that the AI industry has been slow to acknowledge publicly: reducing hallucinations costs time and money, and right now, the competitive pressure to ship faster is winning. Axios reported in June 2025 that AI makers could do more to limit their chatbots' tendency to make things up — but are prioritizing speed and scale instead [1]. The race to deploy, to capture users, to demonstrate capability, is outrunning the race to be reliable.

This trade-off has real consequences at scale. A 2025 survey of Duke University students found that 94 percent believe generative AI's accuracy varies significantly across subjects, and 90 percent want better tools to verify AI-generated information [2]. Students are already navigating this uncertainty — often without adequate guidance from institutions or the platforms themselves. The burden of fact-checking is being quietly transferred from the machine to the user, without much transparency about why.

The legal and medical sectors are feeling this acutely. In high-stakes domains where a fabricated citation or an incorrect drug interaction can have devastating consequences, the hallucination rate of even the best models — which Visual Capitalist's analysis of leading AI systems places at up to 18 percent depending on the task [28] — is simply not acceptable. Enterprise users are discovering that the same tool that writes brilliant marketing copy will, without warning, invent a regulatory requirement that doesn't exist.

Some researchers argue that the framing of hallucination as a solvable bug is itself misleading. Writing for Towards AI, analysts suggest that hallucination may be an intrinsic property of how LLMs generate language — not a defect to be patched, but a feature of probabilistic text generation that can only be managed, not eliminated [22]. This view is gaining traction in academic circles, even as companies continue to promise improvement.

Oxford University researchers working on the reliability of generative models have made progress in measuring and categorizing hallucination types, advancing the field's ability to at least quantify what was previously difficult to define [19]. Measurement, as any scientist will tell you, is the first step toward control. But measurement is not the same as a cure, and the industry's marketing often outpaces its engineering.

AI hallucinations — why language models still make things up - The Road Ahead — Mitigation, Not Magic
The Road Ahead — Mitigation, Not Magic — AI Generated
"The most important feature in any AI system right now is a skeptical human being reading the output."

The Road Ahead — Mitigation, Not Magic

The good news, such as it is, is that the research community is not standing still. Several promising approaches have emerged that, while not eliminating hallucinations, meaningfully reduce their frequency and severity. The most widely adopted is Retrieval-Augmented Generation, or RAG — a technique that grounds a model's responses in real-time retrieved documents rather than relying solely on its internal training [13]. By anchoring the model to verified sources at the moment of response, RAG systems can dramatically reduce the rate of fabrication, particularly for factual queries.

Focused, domain-specific language models represent another avenue. Rather than training a single massive model to know everything about everything, developers are building narrower models trained deeply on specific fields — legal, medical, financial [25]. A model that knows one domain extraordinarily well is less likely to confabulate than one stretched across the breadth of human knowledge. The trade-off is flexibility, but in high-stakes professional contexts, reliability matters more.

OpenAI and other frontier labs are investing heavily in improved evaluation frameworks — better benchmarks that test not just whether a model gives a right answer, but whether it knows when it doesn't know [3]. This metacognitive capacity, the ability to recognize the limits of one's own knowledge, is something humans develop through education and experience. Teaching it to a statistical model is one of the harder problems in the field.

Inside Higher Ed has cautiously suggested that with enough progress in these techniques, AI hallucinations may eventually become rare enough to be manageable rather than endemic [10]. That is a measured hope, not a promise. The Lakera research team notes that the best current models have reduced hallucination rates significantly compared to earlier generations, but the baseline remains troublingly high for critical applications [4].

What is clear is that the solution will not be a single breakthrough. It will be a layered architecture of better training, better retrieval, better evaluation, and — critically — better user education. A 2025 analysis from Getmaxim.ai emphasized that human oversight remains the most reliable hallucination check available today [12]. In other words, the most important feature in any AI system right now is a skeptical human being reading the output. The machine can help you think. It cannot yet be trusted to think alone.

artificial intelligenceAI hallucinationslarge language modelsChatGPTmachine learning
T
The Journaly Crafted by The Journaly — covering technology, culture, and the forces shaping tomorrow.

More in Artificial Intelligence

Share 𝕏 in