Photo by Steve Johnson on Unsplash
You've probably noticed something frustrating lately. That AI tool you loved six months ago? It feels dumber now. Not in an obvious way—it still sounds confident, still formats things nicely. But ask it to solve a problem slightly different from its training data, and it stumbles. You wonder if you're imagining it, or if the company downgraded the model to save money.
You're not imagining it. But the problem isn't what you think.
The Progress That Looked Infinite Suddenly Isn't
From 2018 to 2022, AI followed a predictable script: bigger models, more data, better results. GPT-2 to GPT-3 to GPT-3.5. Each jump felt revolutionary. OpenAI's scaling laws—the mathematical relationship between model size and performance—suggested we could just keep riding this wave indefinitely. Double the parameters, get measurably smarter AI. It was almost boring in its predictability.
Then something odd happened.
Around 2023, despite massive increases in compute and training data, the performance gains started getting... weird. Smaller. Inconsistent. Some tasks improved. Others plateaued. A few actually got worse. Companies started noticing that their flagship models weren't dominating benchmarks the way they used to. They'd announce new models with breathless marketing copy about revolutionary capabilities, and then independent testing would reveal: marginally better, if at all.
This isn't a temporary problem. It's a fundamental challenge that researchers are only now publicly acknowledging, though many saw it coming. The scaling laws that made AI progress feel inevitable are breaking down.
What Happens When You Run Out of Training Data
Here's the uncomfortable truth: we're approaching the limit of publicly available text on the internet. Not tomorrow, but soon enough to matter. Current large language models trained on internet data from up to 2023 or 2024. Every word. Every article. Every social media post that hasn't been deleted. It's all been consumed.
To train bigger, better models, companies need more data. But where does it come from? You can't just scrape the internet harder. The data you haven't used yet is either:
Behind paywalls (academic papers, news archives, books still under copyright). Low quality (random forum posts, spam, AI-generated content). Synthetic (data generated by other AI systems). Ethically murky (personal data, private communications).
OpenAI spent $200 million on processing power for training GPT-4. They couldn't just throw more compute at the old dataset and expect magic. They had to get creative: partnerships for licensed data, synthetic data generation, different training approaches. Each workaround comes with tradeoffs.
Some companies are trying the synthetic route aggressively. Generate training data with existing AI models, then train new models on that synthetic data. Sounds smart in theory. In practice, researchers have discovered that this approach—training on training on training—creates a kind of intellectual inbreeding. The models get stuck in local optima. They lose the messy diversity that makes learning happen.
The Benchmark Gaming Problem Nobody Wants to Admit
There's another reason the progress plateau feels sharper than it is: the benchmarks were gamed to death.
For years, companies measured AI performance against standardized tests. MMLU (Massive Multitask Language Understanding), HellaSwag, Arc, SuperGLUE. These became the scorecard. Every new model release came with updated benchmark scores, and everyone cheered when the numbers went up.
But here's what happens when you optimize for specific benchmarks: you start to overfit to them. Training data leaks into test sets. Companies use test data to guide model development. Models get better at the exact tasks being measured while potentially getting worse at everything else.
When researchers tested newer models on truly novel tasks—things not in the training set or benchmark suite—the improvements looked smaller. Sometimes the models performed worse than older generations. This suggests that newer models aren't actually more capable; they're just better calibrated to pass the tests we use to measure them.
Some researchers have started pointing out that we're measuring the wrong things entirely. A model scoring higher on MMLU doesn't necessarily mean it's more useful for your job. It might just mean it's better at guessing answers in a particular test format.
The Cost Explosion That Nobody Mentions
There's a financial reality underlying all of this: the cost of training frontier AI models has become obscene.
Training GPT-4 reportedly cost somewhere between $50 million and $200 million. For Gemini Ultra, estimates range even higher. These are one-time costs. And the returns are getting smaller. You're spending exponentially more compute to get linearly smaller improvements in performance.
Mathematically, this isn't sustainable. There's a break-even point where the marginal improvement doesn't justify the capital expenditure. No investor wants to fund an AI company that's spending billions to eke out 2% performance gains.
This is why you're seeing a shift in AI strategy across the industry. Instead of racing to train bigger models, companies are focusing on making existing models better: fine-tuning for specific domains, improving inference efficiency, building better interfaces, adding tools and plugins. These generate real business value with a fraction of the cost.
It's less exciting than announcing a bigger model, but it's what actually makes sense at this point in the curve.
What Comes Next (The Honest Version)
Nobody knows if this is a permanent plateau or a temporary wall. The honest answer is somewhere in between.
We'll probably see continued improvements. Different architectures might unlock new capabilities. Better data might come from new sources. But the days of effortless scaling—where bigger automatically meant better—are over. Future progress will be harder, slower, and more expensive per unit of improvement.
The companies that thrive won't be the ones chasing raw model size. They'll be the ones figuring out what AI is actually useful for. Building things people want to use. Creating value instead of just scaling numbers.
Your AI tools might not get spectacularly smarter. But they could get meaningfully more useful. That's a different kind of progress—quieter, less exciting, but ultimately more important.
If you're interested in understanding more about the deeper limitations of current AI systems, you should read about why AI keeps hallucinating about facts it should know, which reveals how these architectural compromises create persistent problems that scaling alone won't fix.

Comments (0)
No comments yet. Be the first to share your thoughts!
Sign in to join the conversation.