Photo by Steve A Johnson on Unsplash
Last year, a lawyer named Steven Schwartz submitted a court brief written with the help of ChatGPT. The problem? Six of the cases he cited were completely fabricated. The AI had invented convincing-sounding legal precedents with official-looking citations. When confronted, ChatGPT apologized politely and explained it had simply "hallucinated" the references. This wasn't a one-time glitch. It's become a defining characteristic of modern large language models—and it's quietly reshaping how we should think about AI reliability.
The Confidence Problem
Here's what makes this peculiar: when GPT-4 makes up a citation, it doesn't sound uncertain. It presents the fake reference with the same confident tone it uses for real information. There's no hedging, no "I think this might be," no epistemic humility. The AI outputs a fabrication as though it were documented fact, complete with author names, publication years, and journal titles that sound entirely plausible.
Researchers call this phenomenon "hallucination," though the term is somewhat misleading. A human hallucinating might see a pink elephant that isn't there. But GPT doesn't see anything—it's generating text based on statistical patterns it learned during training. When those patterns suggest a particular sequence of words (like a fake citation) has high probability, the model outputs it. No internal experience. Just patterns activating patterns.
The real danger emerges when you combine this tendency with human psychology. We trust written citations. They carry authority. When someone quotes a paper, we assume it exists. An entire chain of trust gets built on statistical probability masquerading as fact.
Why This Happens at a Technical Level
Language models work by predicting the next most likely word, token by token. They're trained on massive amounts of text from the internet, academic papers, books, and other sources. During training, the model learns statistical relationships between words and concepts. When you ask it a question, it generates text by repeatedly asking: "Given what came before, what word should come next?"
The problem is that large language models don't actually "know" things the way you might think. They don't retrieve facts from a database. They generate plausible-sounding text based on patterns. If a model has seen thousands of real citations during training, it learns the *format* of a citation—author name, year, title, journal. It learns that citations typically reference real papers. So when asked a question, it generates text that fits this pattern, whether or not the actual paper exists.
Think of it like this: if you trained an AI on thousands of restaurant reviews, it could generate convincing fake reviews of restaurants that don't exist. Not because it's trying to deceive you, but because it learned the statistical structure of restaurant reviews. Citations are just another form of structured text.
This problem gets worse when information is scarce or niche. How AI Learned to Disagree With Itself (And Why That's Making It Smarter) explores how systems are now learning to catch these errors by cross-referencing multiple models. But we're still in early days.
Real-World Consequences Are Already Here
The lawyer's false citations incident might seem like an isolated embarrassment, but it's just the visible tip. Medical students are using ChatGPT to supplement their learning and getting incorrect drug interactions. Researchers are citing AI-generated papers in their work. One academic discovered that a reference he'd seen cited multiple times in a paper turned out to be fabricated—traced back through citations until it hit a dead end.
What's particularly insidious is that detection requires effort. To catch a fake citation, someone has to actually look it up. They have to notice something seems off. Many won't. Studies suggest that people tend to trust information that's formatted authoritatively and presented with confidence. The internet has already trained us to assume that if something is written down, someone fact-checked it.
Some researchers have begun testing AI systems specifically to find where they hallucinate. One study asked GPT-3 about "researchers" who didn't exist and asked it to describe their work. The model cheerfully invented research papers and biographical details. Another study found that models hallucinate more when asked about less common topics—areas where training data is sparse and the model has learned fewer real facts to anchor its outputs.
What Comes Next
The AI research community is working on solutions. Some approaches involve having models explicitly cite their training sources. Others involve building fact-checking mechanisms into the generation process. But these solutions are computationally expensive and not yet standard practice. Right now, the most reliable defense is human verification—which defeats much of the point of using AI to generate answers quickly.
Some researchers argue we need to redesign how language models work fundamentally. Instead of pure generative models, we could build systems that retrieve actual sources and generate text based on them—like a sophisticated search engine combined with a summarizer. This takes more computing power and is slower, but it eliminates hallucination of entire sources.
Others propose transparency requirements: models could explicitly flag when they're uncertain or when information comes from a weaker training signal. They could express probabilities instead of false confidence. But this requires training models differently, and companies building these systems are often more focused on impressive outputs than cautious ones.
The Uncomfortable Truth
The uncomfortable reality is that we don't yet have a killer solution for hallucination in language models. We have workarounds. We have band-aids. We have papers describing the problem in excruciating detail. But the core issue—that these models generate plausible-sounding text without necessarily grounding it in reality—remains baked into their architecture.
For now, the responsibility falls on users. Anyone using AI for research, writing, or decision-making needs to verify outputs independently. Treat them like a first draft from a very confident but sometimes careless colleague. Check the citations. Verify the facts. Don't assume the formatting means something has been checked.
As AI systems become more integrated into professional work, this might seem like a temporary problem that'll get solved. Maybe it will. But it's also possible that hallucination is just the price we pay for systems that are fluent enough to sound authoritative. That's a bargain worth being very careful about.

Comments (0)
No comments yet. Be the first to share your thoughts!
Sign in to join the conversation.