The Iron Age of Compute (DDCU 1/7)
How We Built Cathedrals to Electricity and Called Them Progress
I used to think scale was the answer. Bigger rooms. More racks. More cooling. More contracts with utility companies that couldn’t decide if AI was a crisis or an opportunity. Then I looked at the actual numbers — really looked at them — and realized we’d been building the wrong thing since 1964.
Let me take you back. Not in nostalgia. In autopsy.
The Mainframe Church, 1964–1985
The IBM System/360 launched in 1964. It cost $5 billion to develop — roughly $46 billion in today’s money — and it was, without question, the most audacious bet in the history of computing. Thomas Watson Jr. called it “the biggest risk that IBM has ever taken.”
He was right. He was also building a cathedral.
Mainframes required raised floors. Specialized cooling. Dedicated power substations. Rooms measured in thousands of square feet for machines that, by 2026 standards, have less compute than the chip inside your car’s keyless entry fob. IBM’s first commercial data centers consumed anywhere from 50 kW to 5 MW of power — for systems that, at peak, could process roughly 10 MIPS (million instructions per second).
Your phone processes 15,000 MIPS. Your phone is not climate-controlled in a purpose-built concrete fortress.
The mainframe era established a paradigm that would haunt us for sixty years: compute must be centralized, expensive, controlled, and surrounded by bureaucracy proportional to its power.
This wasn’t irrational. In 1964, compute was genuinely scarce. Centralization was efficiency. Sharing a mainframe across an entire university or corporation made mathematical sense when each transistor represented a miracle of precision manufacturing.
But the paradigm outlived its rationale by decades.
The PC Revolution Didn’t Kill Centralization — It Deepened the Contradiction
When the IBM PC arrived in 1981 and the Mac in 1984, the conventional wisdom was that we’d democratized compute. Decentralization. Power to the people. Every desk its own universe.
Wrong.
What actually happened: the PC created demand for networked services that no desktop could serve alone. Email servers. Database servers. File servers. Application servers. By the mid-1990s, every company that had eliminated its mainframe was building a server room. By 2000, those server rooms had become data centers. Smaller than IBM’s cathedrals, yes. But still centralized. Still dependent on utility power. Still requiring raised floors and precision cooling and specialized staff.
The distributed revolution created a distributed hardware problem, and the industry’s answer was to centralize it again. Just in smaller boxes. Called differently.
In 1998, the average corporate data center consumed 1–2 MW of power for what we would now recognize as trivially modest workloads. Heat was managed with perimeter air handlers and hopes. Power was managed with UPS systems and prayers.
Efficiency, as measured by Power Usage Effectiveness (PUE), hovered around 2.0 — meaning for every watt of useful compute, another watt was wasted on cooling, lighting, and the existential dread of the facilities manager.
The Internet Boom Built the Wrong Infrastructure at Maximum Speed
1996 to 2001. The industry collectively lost its mind. And built a lot of data centers.
Between 1998 and 2001, U.S. investment in data center infrastructure exceeded $45 billion — the equivalent of roughly $80 billion today. Buildings were raised in suburban office parks. Carrier hotels emerged in city centers. Colocation exploded. Every company needed “Internet-scale infrastructure,” a phrase that meant approximately nothing but justified approximately everything.
The assumption baked into every build: these facilities would run at capacity within 18 months. The assumption was wrong approximately 90% of the time. When the dot-com collapse arrived in 2001, something like 40% of all newly built data center space in the U.S. sat dark — unused, half-leased, consuming power for cooling systems that had nothing to cool.
The dot-com boom proved something important: infrastructure built for anticipated demand, at industrial scale, on speculative timelines, is infrastructure that fails expensively and publicly.
The physical plants were fine. The fiber was in the ground. The buildings stood. What was wrong was the model: build massive centralized infrastructure, pray demand fills it, lock capital in concrete for 20-year depreciation cycles.
Nobody learned the lesson. Because the next wave was already coming.
Virtualization and the Cloud: Efficiency Theater
VMware shipped its first x86 virtualization product in 1999. Amazon Web Services launched in 2006. These were genuine revolutions in how compute was consumed — and genuine disasters for how compute was built.
Virtualization allowed one physical server to run dozens of virtual machines. This was extraordinary. It meant data centers could do far more with existing hardware. Utilization rates climbed from the abysmal 5–15% typical of dedicated servers toward 60–70% for well-managed virtualized environments.
AWS meant a startup could launch without owning a single server. The capex-to-opex transformation was real, substantial, and genuinely valuable for customers.
For the underlying infrastructure? Different story.
Amazon’s first data center in Ashburn, Virginia — the so-called “Data Center Alley” that would become the most concentrated compute geography on Earth — consumed roughly 30 MW.
By 2010, Amazon, Google, Microsoft, and Facebook were collectively consuming more than 1 GW globally.
By 2015, that number crossed 10 GW.
By 2024, hyperscaler data centers alone consumed an estimated 45–50 GW worldwide.
The architecture remained identical to 1964’s cathedral. Bigger.
More efficient (PUE had improved to 1.3–1.5 by 2015).
Fundamentally the same structure. Centralized. Grid-dependent.
Permitting-constrained.
Built to last 20 years in landscapes that were changing every 18 months.
The Structural Problems Nobody Was Counting
Here’s what the cloud era obscured: the physical infrastructure was becoming increasingly divorced from the economic and environmental reality around it.
Grid interconnection queues in the United States had grown from manageable to catastrophic. By 2024, the Lawrence Berkeley National Laboratory reported 2,300 GW of projects waiting in U.S. interconnection queues — with only 14% historically reaching commercial operation. Permitting timelines for utility-scale power connections averaged 60+ months.
Data center construction itself required 48–72 months from groundbreaking to operations.
That math is brutal: by the time a traditional hyperscale data center completes construction, the compute it was designed to host has gone through two full GPU generations.
NVIDIA’s H100 — state of the art in 2023 — was already being superseded by the Blackwell architecture before many 2023-vintage facilities came online.
Meanwhile, PJM capacity clearing prices — the wholesale price of grid power capacity in the mid-Atlantic region — surged from $28.92 per MW-day in the 2024–25 planning year to $329.17 per MW-day in 2026–27.
A 10x increase in one year.
Texas enacted SB-6, requiring legislative approval for data centers exceeding 100 MW. Europe implemented permitting cycles of 3–7 years for industrial power connections.
We built an industry on the assumption that grid power was infinite, cheap, and immediate. By 2026, every one of those assumptions had been catastrophically disproven.
The Pre-AI Legacy: What We Inherited
To understand where we need to go, you have to sit with what we inherited.
The pre-AI data center industry — everything from IBM mainframes through the hyperscaler cloud — built infrastructure on four axioms that were true in 1964 and increasingly false by 2024:
One: Compute should be centralized near population centers and corporate headquarters.
Two: Power comes from the grid, and the grid will scale to meet demand.
Three: Buildings should be permanent, purpose-built, and depreciated over 20+ years.
Four: Human staff are required for operations, maintenance, and security.
Each of these axioms made sense in its original context.
Each of them became a strategic liability as AI workloads arrived with their exponential, geography-agnostic, perpetually-accelerating compute demands.
The AI era didn’t just need more data centers. It needed a different architecture entirely.
And the industry, locked into 60 years of infrastructure assumptions, was spectacularly unprepared.
The cathedrals were magnificent.
They just couldn’t survive the reformation.
As we learned in Europe almost 600 years ago.
JF.


