Counting the Atoms · Part I

The Physical Cliff

Everyone's funding bits. Nobody's asking where the atoms come from.

Theo Saville · March 2026

The Gap

On a factory floor in the American Midwest, a machinist is staring at a backlog that stretches to 2028. He's running three shifts. His two best guys are in their late fifties. He turned away a defense contract last month — not because he didn't want the work, but because he physically cannot make the parts fast enough. This is not an anecdote. This is the present tense, March 2026, in hundreds of machine shops across the country.

Here's what the AI conversation is missing: there is a cliff, it is here now, and almost everyone is looking the wrong way.

The thesis fits in one sentence. Capital deploys at software speed. Physical capacity grows at manufacturing speed. That gap is the entire story of what happens next.

A trillion dollars doesn't build a datacenter. A trillion dollars orders a datacenter. Somebody still has to make the parts. And the people who make the parts are booked until 2029.

In 2025, the five largest technology companies collectively committed roughly $350 billion in capital expenditures. In 2026, that number hits $600 billion — a 36% increase. Three-quarters of it is going to AI infrastructure: datacenters, GPUs, networking, power.

These are the largest infrastructure investments in human history. Larger than Apollo. Larger than the interstate highway system. Larger than the Manhattan Project in inflation-adjusted terms. They are being made simultaneously, by competitors, in a race where second place is worthless.

Every serious analysis focuses on chip supply, model architecture, training efficiency, energy policy. The smart money models TSMC fab utilization, NVIDIA's order book, the price of electricity in West Texas.

Nobody is asking where the physical things come from. The answer resets the timeline.


The Number Nobody's Stress-Testing

The AI buildout depends on a stack of physical components — power transformers, gas turbines, switchgear, cooling systems, copper — manufactured by a small number of companies, with multi-year lead times, using materials from concentrated and sometimes sole-source suppliers. The capacity to produce these components cannot scale at the rate capital is being deployed. In several critical cases, it cannot meaningfully scale at all within the relevant timeframe.

This is not a forecast. It is the present tense, March 2026.

Transformer lead times sit at 128–144 weeks. Gas turbine backlogs extend to 2029. Cooling system orders arrive three times faster than manufacturers can fulfill them. Copper is heading into a structural deficit. And the critical path to an operational datacenter runs straight through this bottleneck stack — not through chip supply, not through construction, not through permitting. Through the electrical infrastructure that nobody in Silicon Valley thinks about.

The Datacenter Bottleneck Stack

Layer 0 — Bedrock materials: Grain-oriented electrical steel (GOES) for transformer cores, copper, nickel superalloys. Concentrated production, hard to expand, geological timescales to develop new sources.

Layer 1 — Critical components: Power transformers (128–144 week lead times), gas turbines (3 OEMs globally, backlogs to 2029), switchgear (44+ week lead times), liquid cooling units (supply chain constrained).

Layer 2 — Systems: Utility substations, on-site power plants, cooling infrastructure. Each requires Layer 1 components plus skilled tradespeople to install — welders, pipefitters, electricians whose labor market is already in deficit.

Layer 3 — Datacenter: The building is 12–18 months. The electrical infrastructure is 18–36 months. The critical path runs through electrical infrastructure, not construction.

The building itself — structural steel, concrete, the physical shell — is commodity construction with elastic global supply. What's bottlenecked is everything that makes the building functional: the equipment that delivers power, converts voltage, removes heat, and connects the facility to the grid. Every one of those components traces back to a supply chain that's concentrated, capacity-constrained, and stubbornly slow to expand.

And here's what makes the bottleneck stack uniquely dangerous: it isn't just hardware. It's people. Every transformer that ships needs a team to install, connect, and commission it. Every gas turbine needs specialized welders for the high-temperature piping. Every liquid cooling loop needs technicians to plumb, pressure-test, and certify it. The U.S. is short 427,000 welders by 2028, according to the American Welding Society. It's short 64,000 tool and die makers. The average age of a skilled machinist is 56. Foundry workers, NDT inspectors, composite technicians — every trade that touches this stack is hemorrhaging people faster than it can replace them.

Machinists are the canary. The entire mine is collapsing.

Capital at software speed. Capacity at manufacturing speed. The gap starts here and compounds at every layer.


Transformers: Hand-Wound on a Foundation of Sand

If I had to name the single most important physical bottleneck in the AI infrastructure buildout, it is not GPUs. It is not data cabling. It is power transformers.

A 100MW datacenter campus needs 2–4 large power transformers at the utility substation, plus 20–50 medium-voltage transformers for internal distribution. Without them, you have a very expensive building with no electricity.

Right now, in March 2026, you cannot get them.

Power Transformer Market, 2025

Demand up 119% since 2019 (Wood Mackenzie, Aug 2025)

Generator step-up transformer demand up 274% since 2019

30% supply deficit in the U.S.

Lead times: 128 weeks (power transformers), 144 weeks (generator step-ups)

Prices up 77% since 2019

55% of U.S. distribution transformers are beyond expected service life

128 weeks. Two and a half years. The AI buildout doesn't have two and a half years. By the time your transformer arrives, the model it was ordered for will be three generations obsolete — but the datacenter still won't turn on without it.

Trace this to bedrock. The picture gets worse at every layer.

What's Inside a Transformer

A large power transformer is deceptively simple in concept and brutally difficult to manufacture. Five things matter:

  1. A magnetic core made of grain-oriented electrical steel — a specialty steel whose crystal structure is precisely aligned to minimize energy loss (the industry calls it "GOES")
  2. Copper windings
  3. Insulation — Kraft paper, pressboard, transformer oil
  4. A steel tank
  5. Bushings, tap changers, and cooling systems

The critical constraint is the first item: GOES. Everything else follows from it.

The Grain-Oriented Steel Problem

GOES has a crystal structure precisely aligned to channel magnetic flux with minimal energy loss. It is not optional. You cannot substitute a different steel. You cannot use aluminum. The physics requires it.

Global production runs about 2.5 million tonnes per year. Production is extremely concentrated:

Read that last line again. The United States — which is attempting to build more AI datacenter capacity than the rest of the world combined — has one domestic producer of the steel required for every power transformer in every datacenter. One.

Building new capacity takes 3–5 years. I've watched enough metallurgical projects to know that timeline is optimistic. Making GOES is one of the most demanding processes in commercial production. The silicon content must hit approximately 3%. The steel is rolled, then heat-treated at up to 1,200°C to align the crystal structure — oriented within a few degrees of the rolling direction. Magnetic properties must be uniform across the entire coil. Beijing Shougang started a 90,000 tonne/year line in 2023. It is still ramping in 2026. Nippon Steel launched its Ultra-Loss-Core grade in March 2025. This supply chain moves in half-decades, not quarters.

And the trade policy irony is savage: Section 232 tariffs, plus the newer "Liberation Day" tariffs, are raising costs for U.S. transformer manufacturers who need to import GOES — because Cleveland-Cliffs alone cannot supply enough. Trade policy designed to protect American manufacturing is actively strangling the transformer production that American datacenters need. I've talked to a utility procurement executive who described the situation in two words: "policy whiplash." He's ordering transformers for a substation that won't energize until 2029. The tariff regime has changed three times since he placed the order.

The Hand-Winding Problem

Even if you solved the steel supply tomorrow, you'd hit the next wall. Large power transformers are hand-wound.

Sit with that. The most important physical component in a trillion-dollar AI infrastructure buildout — the thing without which no datacenter gets electricity — is assembled by human beings, winding copper by hand.

Per Fastmarkets (October 2025): "transformers are hand-wound in a process that cannot be automated." A large power transformer requires weeks of skilled manual labor for winding alone, plus extensive factory testing — impulse testing, thermal testing, partial discharge testing — that takes 2–4 weeks per unit and cannot be parallelized.

This process resists automation in a fundamental way. The variation between transformer designs, the precision required in winding copper around the cores, the quality assurance at each layer — you need skilled people. And the skilled people are aging out. The median transformer winding technician in the U.S. is in their mid-fifties. Their replacements are not in a training pipeline, because there is no training pipeline. The same demographic collapse hitting machinists, welders, and tool and die makers is hitting the tiny workforce that winds the transformers the AI industry depends on.

There is roughly $1.8 billion in announced North American transformer manufacturing expansion since 2023 — Hitachi Energy building the largest U.S. plant in South Boston, Virginia ($457M, target 2028), Siemens Energy in Charlotte, NC ($150M, early 2027), Eaton in South Carolina ($340M, 2027). Real money. But these investments won't deliver meaningful additional capacity for 2–3 years. And they don't solve the GOES concentration, the tariff squeeze, or the labor crisis.

The capital is being deployed. The capacity will follow. Just not at the speed anyone assumes.


The Rest of the Stack

Transformers are the worst bottleneck. They are not the only one. Every layer of this stack tells the same story: capital at software speed, capacity at manufacturing speed. The gap widens as you go.

Gas turbines. Hyperscalers are bypassing the grid — interconnection queues run 5+ years — and building their own power plants. The problem: three companies on Earth make heavy-duty gas turbines. Three. GE Vernova, Siemens Energy, Mitsubishi Power. GE Vernova's order book exceeds $73 billion with an 80 GW backlog (Q3 2025). Siemens Energy's backlog hit a record €131 billion (FY2025). Deliveries stretch to 2029. The Stargate project alone is procuring 4.5 gigawatts of turbines — the generating capacity of a mid-sized country. Nobody else is entering this market. Developing a new heavy-frame turbine costs billions and takes a decade. The oligopoly is permanent.

A datacenter project manager I know describes it this way: "We have the site, the permits, the capital, and the customer. We're waiting on a turbine. The turbine is waiting on a forging. The forging is waiting on a nickel superalloy pour. That's three queues deep before we generate our first watt." Each of those queues depends on welders who can work to aerospace-grade specs, NDT inspectors who can certify the welds, and foundry workers who pour the superalloy castings. Every one of those trades is in deficit.

Cooling. AI racks run at 120+ kW per rack — versus 30 kW for traditional servers. Air cooling is physically impossible at that density. Every new AI datacenter requires direct liquid cooling — precision-plumbed loops running through every rack. Vertiv, the dominant supplier, tells the story in one number: backlog went from $8.5 billion to $15 billion in six months, with a book-to-bill ratio of 2.9x. Orders arrive nearly three times faster than they ship. These cooling systems need copper tubing, brazed by certified welders, pressure-tested by inspectors, and installed by technicians who understand both thermal dynamics and plumbing codes. The hardware bottleneck and the labor bottleneck are the same bottleneck.

Copper. It is in everything — transformer windings, power distribution, cable runs, cooling components, generator windings. A single 100MW campus uses 2,700 tonnes. The market is heading into a 304,000-tonne deficit in 2025, with datacenter demand projected to hit 1.1 million tonnes per year by 2030. New mines take 10+ years to develop. The copper for the datacenters coming online in 2030 needed to be discovered and permitted years ago. Much of it was not.

Switchgear. Lead times past 44 weeks. Prices up 50%. Every substation, every distribution panel, every point where voltage is switched or protected requires it. Not glamorous. Not optional.

The pattern repeats at every layer: concentrated suppliers, multi-year lead times, physical processes that resist rapid scaling, a workforce that is aging out with no replacement generation. The gap between capital deployment speed and physical capacity is not closing. It is compounding.


Why Nobody Sees This

Raise this at conferences, investor meetings, dinners. The reaction is always the same: a flicker of surprise, then "well, they'll figure it out." They meaning someone else. Someone downstream. Someone whose problem this is.

Three dynamics keep this invisible to the people making the biggest capital allocation decisions of the decade.

First, the AI investment thesis lives entirely on the digital side. The models that drive hundred-billion-dollar allocation decisions model chip supply, training efficiency, and revenue per inference call. They do not model transformer lead times, GOES allocation, welder availability, or how fast cooling units actually ship. The assumption — usually implicit, sometimes embarrassingly explicit — is that once you commit the capital, the physical infrastructure just appears. It does not. I have watched this assumption destroy production schedules for a decade.

Second, the bottleneck manifests as delay, not failure. No hyperscaler announces "we couldn't build our datacenter because we couldn't get transformers." They announce a six-month delay. The transformer wait gets buried under "construction timeline adjustments." The constraint is real but structurally invisible to anyone not inside the supply chain. By the time it surfaces in an earnings call, it has been euphemized beyond recognition. A machinist at a transformer manufacturer told me last fall: "We're running three shifts and we're still 18 months behind. But the customer gets told it's a 'supply chain adjustment.' Nobody says we just don't have enough hands."

Third — and this is what keeps me up at night — the physical world runs on a fundamentally different clock. In AI, a year is a geological epoch. Models double in capability. Architectures get reinvented. Startups go from founding to billion-dollar valuations. But a GOES production line takes 3–5 years to build. A gas turbine factory takes years to ramp. A copper mine takes a decade. Training a welder to aerospace specs takes four years. You cannot compress the apprenticeship because a human nervous system learns at the speed it learns. The physical world does not care about your roadmap. Its timescales are geological, metallurgical, thermodynamic, and biological. They don't compress because your stock price needs them to.

This mismatch — capital at software speed, capacity at manufacturing speed — is not a bug that gets patched. It is the governing constraint of the entire AI buildout. And almost nobody modeling the future of AI has it in their model.


The Cliff

Here is the full picture. It is March 2026.

The AI industry is deploying $600 billion in capital against an infrastructure supply chain that requires: power transformers (128–144 week lead times, 30% U.S. supply deficit, dependent on GOES from a single domestic producer), gas turbines (three manufacturers globally, order books exceeding $200 billion, deliveries stretching to 2029), switchgear (44+ week lead times, 50% price increases), cooling systems (orders arriving 3x faster than the dominant manufacturer can ship), copper (heading into a 304,000-tonne deficit with 10+ year mine development cycles), and a skilled industrial workforce that is short nearly half a million welders, tens of thousands of tool and die makers, and an entire generation of machinists, foundry workers, NDT inspectors, and composite technicians.

Every one of these constraints traces to physical processes that resist rapid scaling: specialty metallurgy that takes years to expand, hand-wound assembly that defies automation, oligopolies with multi-year delivery queues, geological resources with decade-long development timescales, and human expertise that takes years to develop and is retiring faster than it is being replaced.

The capital is being committed at software speed. The physical capacity grows at manufacturing speed. The gap between these two rates is the cliff. And we are standing on it right now.

By the time you finish reading this four-part series, hundreds of millions more dollars will have been committed to datacenter builds that depend on transformers that haven't been wound yet, turbines that haven't been cast yet, copper that hasn't been mined yet, and workers who haven't been trained yet. The orders are accelerating. The supply chains are not. The workforce is shrinking.

This doesn't mean the AI buildout fails. It means something more consequential: the buildout's timeline is not set by how fast capital deploys, or how fast GPUs ship, or how fast models improve. It is set by how fast someone can hand-wind a power transformer. How fast GE Vernova can deliver against a $73 billion backlog. How fast Vertiv can ship cooling units at 2.9x book-to-bill. How fast copper comes out of the ground. How fast you can train a welder to certify superalloy pipe joints at 1,400°F.

The intelligence explosion has a supply chain. And the supply chain has physics.

The people building the future of intelligence are discovering — right now, in 2026 — that the future of intelligence is gated by the past of manufacturing. By hand-wound copper, by specialty steel from a single American mill, by three companies who make turbines and nobody else who can learn fast enough, and by a generation of tradespeople who are retiring and taking their knowledge with them.

If the rate-limiter for AI is not compute but physical infrastructure, the implications reshape the competitive landscape of the entire AI race — which nations can build, which companies can scale, which bets pay off and which ones stall behind a transformer that won't arrive until 2028, waiting for a welder who retired in 2027.

It goes deeper than anyone expects. The supply chains have supply chains, and those have supply chains. The Manhattan Project's bottleneck was enriching uranium. The AI revolution's bottleneck is manufacturing the machines that manufacture the machines that manufacture the chips. Each layer adds years. It's bottlenecks all the way down.

Part II: Turtles All the Way Down →

Sources & Data

Theo Saville is a manufacturing, mechanical, and robotics engineer. He is CEO and co-founder of CloudNC, a company building AI for CNC machining, Honorary Professor of Engineering at the University of Warwick, and an MIT Technology Review Innovator Under 35. Before that, he worked in additive manufacturing research and defence engineering. He has spent the last decade at the intersection of AI and physical manufacturing.