The Race to Discover if AI will Scale
23 December 2024
For years, the tech industry has operated under a near-religious belief in the power of scale. If a model wasn’t working quite as intended, the answer was often to make it bigger. Add more GPUs, train on more data, and—presto!—the promise of artificial intelligence could march forward, solving problems from cancer detection to self-driving cars.
Yet, as headlines about megaprojects like data centers housing a million GPUs flash across the news, there’s a growing question in Silicon Valley and beyond: is scaling up still the answer, or are we just chasing diminishing returns?
This question isn’t academic. In the last decade, AI's scaling law—essentially the idea that performance improves predictably with larger models and more training data—has been the industry’s guiding star. Companies like OpenAI, Google, and Meta poured billions into creating ever-larger systems, and the strategy worked, up to a point. Models like GPT-4 dazzled the world with their ability to write poetry, generate code, and even pass professional exams. But then came the first real signs of trouble.
Take OpenAI’s recent release, Orion. Unlike its predecessors, which showed exponential leaps in capability, Orion’s improvements over GPT-4 have been...underwhelming. Early adopters note that while it’s faster and slightly better at handling edge cases, it doesn’t feel revolutionary. If anything, it feels like squeezing the last drops of juice from an already dry fruit. The same pattern is playing out across the industry: as models get larger, the gains shrink while the costs explode.
This shift has profound implications. First, there’s the sheer expense. Training a top-tier AI model costs tens—sometimes hundreds—of millions of dollars. Running these models at scale adds operational costs in the billions. When these costs no longer translate to groundbreaking improvements, investors start to ask tough questions. What happens to the companies betting their futures on this approach? What happens to Nvidia, whose explosive growth depends on the insatiable demand for GPUs? The AI gold rush that has buoyed tech stocks for the past several years may finally hit its first major plateau.
Beyond the financials lies another pressing issue: the environmental toll. Training one large AI model requires as much energy as powering tens of thousands of homes for a year. Multiply that by the growing number of megaprojects, and you’re looking at a significant strain on global energy resources. All of this comes at a time when the world is grappling with climate change, making it increasingly difficult to justify these energy-intensive pursuits without demonstrable societal benefits.
So, what happens next? If the scaling law is no longer the guaranteed path to better AI, the industry will need to pivot. One emerging solution involves focusing not on making models larger, but on making them smarter. Techniques like “test-time compute,” which optimize the inference process rather than the training phase, could improve efficiency without requiring massive hardware investments. Smaller, task-specific models are another promising avenue. These systems, trained to excel in niche areas, may not be flashy, but they are practical and accessible to a wider range of developers.
Interestingly, this shift could democratize AI development. For years, the dominance of scaling meant only the richest companies could afford to compete in the AI race. If smaller, leaner models become the norm, we might see a resurgence of innovation from startups and independent researchers—groups historically known for their creativity and agility.
Regulators, too, are likely to step in as the stakes rise. The environmental costs of AI are becoming impossible to ignore, and public scrutiny is intensifying. We’ve already seen governments call for stricter oversight of AI’s ethical and societal impacts; it’s only a matter of time before energy usage enters the conversation. Expect debates over sustainability standards for AI projects, similar to those faced by the cryptocurrency industry.
In the next six months, we’ll likely see a recalibration. Companies that have long thrived on scale will begin to hedge their bets, exploring alternatives that balance performance with sustainability. Investors will demand clearer answers about returns on investment. And a public that once marveled at AI’s potential may start asking more pointed questions about what all this technology is really for—and who it benefits.
The era of infinite scaling may be ending, but that’s not necessarily a bad thing. Progress isn’t always about growing bigger; sometimes, it’s about growing smarter. The choices made in the coming months will shape not just the future of AI, but its place in society—and whether that place is one we can afford, in every sense of the word.