Photo by Igor Omilaev on Unsplash
Last year, a radiologist at a major teaching hospital showed me something that made my stomach drop. An AI model trained to detect tumors in chest X-rays had confidently identified a malignant growth in a completely normal scan. When asked to explain its reasoning, the system pointed to an area that was literally just empty space—air in the lungs. The model wasn't guessing. It was hallucinating with absolute certainty.
This isn't an isolated incident. Across healthcare systems, law firms, and financial institutions, AI models are doing something unsettling: they're generating plausible-sounding information that has no basis in reality. And because these outputs are presented with such conviction, they're often believed before anyone bothers to verify them.
The Hallucination Problem Is Worse Than We Thought
When researchers at Johns Hopkins tested several large language models on medical knowledge, the results were genuinely disturbing. Claude 3 Opus, one of the most advanced models available, confidently provided "medical advice" that contradicted established clinical guidelines. GPT-4, despite its sophistication, invented drug interactions that don't exist. These aren't models that are slightly confused—they're systems that generate false information with the kind of confidence usually reserved for certainty.
But here's what's really happening: these models aren't being careless. They're doing exactly what they were designed to do. Large language models predict the next word in a sequence based on statistical patterns in training data. When a model encounters a question it hasn't seen before, it doesn't know it's stumped. It just keeps predicting the next most statistically likely word. Sometimes that leads to coherent, accurate responses. Sometimes it leads to complete nonsense presented as fact.
The problem becomes acute in specialized domains like medicine. A language model trained on internet text has seen thousands of real medical facts, but also countless false claims, outdated information, and pseudoscience. When generating medical content, it's essentially rolling dice with the entire corpus of human medical writing as the outcome space. The model has no internal mechanism to distinguish between verified facts and garbage.
Hospitals Know About This. They're Deploying AI Anyway.
What's happening now is that healthcare institutions are moving forward with AI deployment despite these documented failures. A survey by the American College of Radiology found that 82% of hospital systems had either implemented or were planning to implement AI diagnostic tools. Meanwhile, studies continue to show these tools generating dangerous hallucinations.
This isn't recklessness exactly—it's more complicated than that. Hospital administrators face genuine pressure. They need to reduce costs. They're understaffed. A tool that's correct 90% of the time still seems like an improvement when radiologists are burned out and backlogs are growing. The problem is that the 10% it gets wrong isn't random. It's often confident and plausible, making it harder to catch than an obviously broken system.
One teaching hospital I spoke with implemented an AI triage system for emergency departments. The system sometimes hallucinated patient histories, generating fictional medical records that didn't match any actual information. Nurses caught most of the errors, but not all. When I asked the hospital's CTO why they'd deployed a system with known hallucination issues, he didn't defend the decision. He just said: "We're aware of the limitations. We're hoping the next version is better." That's not a strategy. That's a prayer.
Why Detection Is Harder Than You'd Think
The insidious part of AI hallucination is how difficult it is to catch. If a model generated complete gibberish, we'd reject it immediately. But these systems generate content that sounds right. It has the right structure, the right vocabulary, the right tone. It's just... false.
Consider a real example: a language model confidently explained that acetaminophen should not be used in patients with fever over 103 degrees Fahrenheit because it causes "hepatic binding complications." This is entirely fabricated. But it sounds medical. It has specifics. Someone skimming quickly might believe it.
Verification requires expertise. You need someone who actually knows medicine to check the AI's work. But the entire reason hospitals want AI in the first place is to reduce the burden on experts. So we've created a system where the institution most motivated to use AI is also the institution least able to verify its outputs.
This is why AI models learning to construct convincing falsehoods is becoming such a critical problem. It's not just a technical issue to be solved in the next model version. It's a structural problem: systems that generate high-quality-looking content without any internal mechanism to verify truthfulness.
What Actually Needs to Happen
Some hospitals are taking hallucination seriously. They're implementing robust verification workflows. At one major medical center, every AI-generated diagnosis is reviewed by a human radiologist before it ever reaches the patient's chart. This completely defeats the efficiency argument—you're essentially just using AI as a pre-screening tool—but it actually works. No hallucinations reach patients.
The cost is significant. Those human reviews take time. But the alternative is worse: deploying systems that generate dangerous false information and hoping to catch the failures before they harm someone.
Real progress requires AI developers to build systems that can express uncertainty. Instead of predicting the next word with 97% confidence, models should flag when they're operating outside their training domain. They should refuse to answer medical questions if they haven't been specifically trained on verified medical data. This is technically possible. It's just not the default behavior of current systems.
Until that happens, we're in an awkward transition period. We have powerful AI tools that generate plausible-sounding outputs at scale, combined with institutional pressure to deploy them, combined with limited ability to verify their accuracy. It's a recipe for systematic false information embedded into critical systems.
The radiologist I mentioned earlier? She's still using the AI system. But she treats it with extreme skepticism, verifies everything, and has noticed patterns in where it tends to hallucinate. That's not actually a use case. That's just an expensive way to generate busywork for an expert who already knows how to read X-rays without the AI.
Until hallucination stops being a feature and becomes a bug we actually solve, AI in healthcare will remain a tool that requires constant supervision. And if you need constant supervision, you're not really saving anyone's time.

Comments (0)
No comments yet. Be the first to share your thoughts!
Sign in to join the conversation.