On the floor, this makes good sense: extra fashions, information and computing energy. However a rising refrain of researchers and technologists is starting to ask a pointed query—what if we’re getting ready for a compute-heavy world that will by no means arrive?
Joe Tsai, chairman of Alibaba, raised this concern final week. “I begin to see the start of some type of bubble… I’m nonetheless astounded by the kind of numbers which might be being thrown round within the US about investing in AI,” he stated at a convention in Hong Kong (shorturl.at/CkHGE).
Additionally Learn: The nice AI reboot: Educators, techies and leaders all have to adapt quick
This isn’t idle scepticism. Current breakthroughs in GenAI have hinted at a shock development: higher fashions and fewer computing. DeepSeek, a Chinese language mannequin, not too long ago surprised observers with its capabilities and effectivity. It achieved outcomes akin to GPT-4 on many benchmarks whereas working on considerably much less {hardware}.
Across the identical time, Qwen, a basis mannequin from Alibaba, demonstrated related effectivity; it educated utilizing optimization methods that dramatically diminished vitality consumption and dependency on high-end infrastructure. Different GenAI purveyors have responded with equally ‘lean’ fashions. Tsai’s musings are alarming, however probably additionally strategic, given its Qwen ambitions.
Nonetheless, there may be an underlying shift in how AI progress is being achieved. The preliminary wave of GenAI progress—spearheaded by OpenAI, Anthropic and others—relied on brute drive: scaling up parameters, information and computing. This led to fashions that have been spectacular however power-hungry. Coaching GPT-3 required large quantities of computing. GPT-4, by all accounts, was much more demanding. The tech business’s response was to construct {hardware} for this trajectory. If success relied on scale, then hyperscale was the best way ahead.
Additionally Learn: Keep forward of the curve within the age of AI: Be extra human
This has profound implications for the present information centre growth. Amazon, Alphabet and Meta alone plan to spend as much as $240 billion this 12 months. A lot of that is channelled into GPU clusters, power-hungry cooling techniques and grid-scale vitality contracts. Whereas these are helpful for present workloads, their utility might decline quicker than anticipated.
There’s historic precedent for this sort of overinvestment. In the course of the early 2000s’ dotcom bubble, telecom firms laid down huge fibre-optic cable networks, anticipating insatiable web demand. The demand did arrive—however years later. Within the interim, many firms went bankrupt and their infrastructure was mothballed or bought low cost. The one winner was Indian IT outsourcing—when Western firms realized they might now cheaply transfer work to Indian employees utilizing this overcapacity, inflicting the preliminary IT outsourcing growth. Solely a lot later did content material streaming, cloud computing and smartphones put that overcapacity to make use of.
The danger with AI information centres is that they’re extra specialised and fewer forgiving. A fibre-optic cable can be utilized over a protracted interval; GPUs, alternatively, have a brief shelf life. An information centre optimized for dense GPU workloads might not adapt rapidly to lighter and extra broadly distributed AI deployments. Energy contracts signed in anticipation of 2025’s peak utilization may turn into liabilities if fashions in 2026 want solely a fraction of that capability.
Additionally Learn: The iPhone wants a brand new technique to salvage Apple Intelligence
In fact, tech giants are usually not naive. Firms like Microsoft have scaled again a few of their information centre lease contracts (shorturl.at/4XbhH). Nonetheless, some argue that even when mannequin coaching turns into extra environment friendly, true AI scale—serving billions of customers—will nonetheless require huge infrastructure.
There’s additionally an environmental angle to think about. The AI business’s starvation for computing has alarmed local weather researchers. Extra environment friendly fashions are usually not only a cost-saving measure, however a local weather crucial. If tech firms are critical about sustainability, they have to pivot in the direction of much less resource-intensive approaches.
So why does the spending spree proceed? Partly, it’s inertia. Budgets and build-outs deliberate one or two years in the past are solely now coming on-line. There’s additionally a aggressive signalling recreation—in case your rival is spending billions on AI, you don’t need to be seen standing nonetheless.
Additionally Learn: AI chips: India should get its fundamentals proper to satisfy this nice catch-up problem
However most of all, it’s a guess that AI will stay compute-intensive and that infrastructure possession will supply a strategic benefit. That guess might repay within the quick time period, particularly for firms with proprietary basis fashions or working large inference platforms. However finally, it’s doable that probably the most beneficial AI firms gained’t be these with the most important information centres, however these with the neatest algorithms.
For now, bulldozers are digging, chips are arriving and the ability meters are spinning. However the true frontier of GenAI might not be in silicon however in code—in coaching strategies, mannequin design and intelligent optimizations that do extra with much less.
If that’s the case, then at present’s information centre growth may resemble a race to construct superhighways solely to find that most individuals desire flying.
The writer is co-founder of Siana Capital, a enterprise fund supervisor.












