The Cathedral Problem: Why Hyperscale Is Architecturally Obsolete
The bigs are spending hundreds of billions building hyperscale data centers. Most of that investment will be structurally misaligned with the AI workloads it was designed to deliver.
Microsoft, Google, Amazon, and Meta are spending hundreds of billions building hyperscale data centers. Most of that investment will be structurally misaligned with the AI workloads it was designed to serve before it comes fully online. This is not a minor inefficiency. It is a foundational architectural mismatch.
I want to be precise about what I mean when I say hyperscale data centers are architecturally obsolete. I do not mean they will stop operating or become worthless. I mean that the design assumptions embedded in the facilities being built right now — the power density per rack, the cooling architecture, the modularity, the time-to-deployment — are calibrated for a world of cloud computing workloads that is being rapidly superseded by a world of AI inference workloads. Those are fundamentally different physical requirements.
What a Hyperscale Facility Is Actually Designed For
A hyperscale data center is an extraordinary engineering achievement. We are talking about facilities that can house millions of servers, deliver gigawatts of power, maintain sub-millisecond internal latency, and operate at 99.999% uptime. The largest facilities — Northern Virginia, Singapore, Dublin, Tokyo — represent decades of accumulated infrastructure development, billions in sunk cost, and engineering excellence at genuine scale.
They were designed for traditional cloud workloads. Email. Storage. CRM systems. Web applications. Virtual machines. The kind of computing where a rack might draw 8-12 kilowatts of power, where air cooling is sufficient, and where you can pack thousands of general-purpose servers into a standard facility footprint.
Between 2021 and 2024, average data center rack power densities rose from 8 kW to 17 kW. That was already pushing the limits of air cooling in standard facilities. By early 2026, AI-driven racks frequently exceed 50 kW — and Nvidia’s Blackwell architecture is pushing toward racks that draw 120 kW, with the next generation aiming for 200 kW and beyond.
8 kW
Average rack density 2021
17 kW
Average rack density 2024
50+ kW
AI racks in early 2026
200 kW
Next-gen AI rack targets
A facility designed for 8-17 kW racks cannot simply be retrofitted to serve 120-200 kW racks. The power delivery infrastructure, the cooling systems, the structural load requirements — everything changes at those densities. You are not upgrading a car. You are trying to turn a highway into an airport runway.
The Construction Timeline Problem
Here is the second dimension of the cathedral problem, and in some ways it is more damaging than the density mismatch. Hyperscale data centers take a long time to build. From site selection to full operational status, you are looking at 4-7 years under normal conditions. Under current conditions — with transformer lead times exceeding 160 weeks and switchgear timelines similarly extended — you could be looking at 7-10 years for a large facility with complex power requirements.
The AI model generations cycle is currently running at roughly 12-18 months. By the time a hyperscale facility designed for today’s AI workloads comes online, the architectural requirements will have shifted at least two full generations. The Nvidia H100 was the dominant AI training chip when many of the facilities currently under construction were designed. By the time those facilities open, the industry will be on H200, Blackwell, or Rubin — each generation requiring fundamentally different power and cooling configurations.
“A hyperscale facility is a cathedral. Cathedrals take decades to build, are designed for a specific liturgy, and are very hard to adapt when the religion changes.”
This is not a hypothetical risk. Microsoft, one of the most sophisticated capital allocators in technology history, already canceled or deferred more than 2 GW of data center capacity in late 2024 and early 2025, according to TD Cowen’s analysis. The explicit reason: the facilities were designed around training workloads for OpenAI models. When that relationship evolved, the facilities became misaligned with actual demand. Two gigawatts of stranded capacity. That is not rounding error.
The Cooling Problem Is Not Solvable With Air
Standard air cooling systems work adequately up to roughly 30 kW per rack. Above that threshold, air cooling requires so much airflow that it consumes significant additional power itself, creates hotspots regardless of airflow volume, and begins to compromise the physical reliability of the hardware it is trying to cool. At 50 kW per rack — which is already common in AI deployments — air cooling is marginal. At 120 kW, it is non-functional.
AI data centers at the frontier are now deploying liquid cooling: direct-to-chip cooling where coolant pipes route directly to GPU heatsinks, rear-door heat exchangers, and immersion cooling where servers are submerged in dielectric fluid. Each of these approaches requires fundamentally different facility design. You cannot retrofit a standard air-cooled hyperscale facility for immersion cooling without effectively rebuilding it.
This means that a significant portion of the hyperscale capacity being financed and constructed right now will be structurally unable to serve the AI workloads that will dominate computing demand by 2028-2030. The facilities will not be worthless — traditional cloud workloads are not disappearing. But for the highest-value AI use cases, they will be the wrong tool for the job.
Why Modularity Is the Structural Answer
If the cathedral is architecturally locked to its original design, the answer is not a bigger cathedral. The answer is a tent — or more precisely, a deployable, reconfigurable modular infrastructure that can be designed, built, and commissioned in parallel with the hardware evolution, co-located with power sources rather than constrained by grid interconnection queues, and upgraded or reconfigured as workload requirements evolve.
A modular data center can be deployed in 6-18 months from commitment to operational status. It can be designed around the specific power and cooling requirements of the GPU generation being deployed, rather than averaged across a multi-decade facility lifetime. It can be positioned where the power is, rather than where the real estate is cheapest. And when the next GPU generation arrives with different requirements, the modular units can be retired or reconfigured without stranding billions in fixed infrastructure.
Crusoe Energy and Energy Vault announced exactly this approach in February 2026 — a partnership to deploy modular “powered shell” data centers designed for high-density AI compute, co-located with stranded energy resources. Eaton and Siemens Energy announced a similar collaboration, combining modular power generation with modular IT infrastructure designed for parallel construction and deployment.
The market is arriving at the same conclusion independently. The cathedral model is breaking. The tent model is emerging. The only question is who builds enough tents fast enough to capture the demand that hyperscale is structurally incapable of serving.
The key insight: Hyperscale facilities are not being abandoned — they are being outpaced by a workload revolution they were not designed to serve. The fastest-growing category of AI compute demand (high-density inference, edge AI, distributed training) requires exactly the characteristics that hyperscale cannot provide: speed of deployment, flexibility of density, and proximity to power rather than population centers.



Thanks for sharing this Jiri !