Chapter 1
The AI Infrastructure Cycle
Chapter 1: The AI Infrastructure Cycle
1.1 Scale of the Buildout
The AI infrastructure buildout is the largest concentrated capital investment cycle in the history of the technology industry, and likely in the history of private enterprise. The numbers are not subtle.
As of Q1 2026 earnings, the four largest hyperscalers (Amazon, Alphabet, Microsoft, Meta) collectively plan to spend approximately $725 billion on capital expenditures in calendar year 2026, up 77% from 2025’s record $410 billion 1. Adding Oracle ($50B target) and the three largest Chinese cloud providers (ByteDance, Tencent, Alibaba, Baidu), TrendForce estimates total top-nine CSP capex at approximately $830 billion for 2026 2.
Company-by-company 2026 capex guidance (calendar year, most recent guidance):
| Company | 2026 Capex Guidance | 2025 Actual/Est. | YoY Growth | Source |
|---|---|---|---|---|
| Amazon (AWS + other) | ~$200B | ~$125B | ~60% | 3 |
| Alphabet/Google | $180-190B | ~$91B | ~100%+ | 1 |
| Microsoft | ~$190B | ~$90B | ~110% | 14 |
| Meta | $125-145B | ~$72B | ~85% | 12 |
| Oracle | ~$50B | ~$20B | ~150% | 5 |
Important caveats on these headline figures: Amazon’s $200B includes non-data-center capex such as Project Kuiper (satellite constellation) and the Globalstar acquisition 4. Microsoft’s $190B includes approximately $25B attributed to “higher component pricing,” primarily memory chips 146. Alphabet raised its guidance range partly due to the Intersect data center acquisition, which closed in March 2026 4. Two-thirds of Microsoft’s most recent quarterly capex went to short-lived assets (primarily GPUs and CPUs), meaning a significant portion of this spending must be renewed every 3-5 years 4. This is not a one-time buildout but a recurring cost structure.
Capital intensity and financing:
Capital intensity (capex as % of revenue) has reached 45-57% at some hyperscalers, ratios historically associated with utilities or industrial companies, not technology firms 7. This is not a cyclical anomaly. As Doug O’Laughlin argues in Fabricated Knowledge, the era of zero marginal cost internet software is ending; AI brings back heavy marginal costs in power, silicon, and cooling, transforming technology companies into capital-intensive industrials 40. The competitive logic resembles a dollar auction: every hyperscaler CEO acknowledges the risk of overspending, yet the cost of being excluded from the next computing platform exceeds the cost of overinvestment 40. Approximately 75% of aggregate hyperscaler capex in 2026 is directed at AI-related infrastructure. AI-specific spending for the top five US hyperscalers is estimated at roughly $450B after excluding non-AI capital programs 7. Goldman Sachs projects cumulative hyperscaler capex from 2025-2027 will reach $1.15 trillion, more than double the $477B spent from 2022-2024 8. Bank of America analyst Vivek Arya projects global semiconductor sales will surpass $1 trillion in 2026, a milestone previously not expected until 2030, and characterizes the industry as being at the “midpoint of a decade-long transformation” 41. Since GPT-4’s release (March 2023), combined Big Five capex has grown at an annualized rate of 72% (90% CI: 66-78%), per Epoch AI’s analysis of SEC filings using structured XBRL tags rather than company-reported aggregate figures 9.
Capex now exceeds internal cash generation at several hyperscalers. Amazon is projected to have negative free cash flow of $17-28B in 2026 (estimates from Morgan Stanley and Bank of America respectively). Alphabet’s free cash flow is projected to fall approximately 90% to roughly $8.2B in 2026 from $73.3B in 2025 (Pivotal Research estimate). Hyperscalers raised $108B in debt during 2025 alone, and projections from Morgan Stanley and JP Morgan suggest the technology sector may need to issue $1.5 trillion in new debt over the coming years to finance construction 68.
1.2 Where the Money Goes
A GPU cluster is not just GPUs. The bill of materials for a modern AI data center spans dozens of component categories. A rough decomposition of hyperscaler capex (varies by company and project):
- Servers and GPUs: roughly 60-65% of total IT capex. NVIDIA GPUs (or AMD/custom ASICs), plus the server chassis, motherboards, power delivery, and memory (HBM).
- Networking: roughly 10-15%. Switches, optical transceivers, cables, NICs/DPUs.
- Storage: roughly 5%. SSDs, HDDs, storage arrays.
- Power infrastructure: roughly 10-15% of facility capex. Transformers, switchgear, UPS, PDUs, generators, substations.
- Cooling: roughly 5-10% of facility capex. CRAC units, liquid cooling loops, chillers, cooling towers.
- Building and civil works: roughly 10-15% of facility capex. Site prep, concrete, steel, construction labor.
- Land and interconnection: Variable. Increasingly significant as powered land becomes scarce.
Capex allocation waterfall (see Section 1.2). The following table translates the Big Four hyperscaler capex ($725B) into approximate dollar flows to each supply chain layer analyzed in this report. These are order-of-magnitude estimates, not audited allocations; they are provided so that the reader can assess each chapter’s market sizing in context. Total spending including Oracle and Chinese CSPs is higher (~$830B).
| Layer | Approx. % of total capex | Approx. $ (of $725B) | Key chapters |
|---|---|---|---|
| Servers, GPUs, CPUs, HBM | 45-50% | $350-390B | Ch. 6, 8, 18 |
| Advanced packaging (CoWoS, substrates) | 3-5% | $25-40B | Ch. 9 |
| Networking silicon, switches, DPUs | 5-8% | $40-60B | Ch. 10 |
| Optical transceivers, fiber, connectors | 4-7% | $30-55B | Ch. 11, 12 |
| Semiconductor equipment (indirect, via wafer pricing) | 2-3% | $15-25B | Ch. 4 |
| Semiconductor materials (indirect, via wafer pricing) | 1-2% | $8-15B | Ch. 2, 3 |
| Power infrastructure (transformers, switchgear, UPS) | 5-8% | $40-60B | Ch. 13, 14 |
| Cooling (liquid cooling, HVAC, chillers) | 3-5% | $25-40B | Ch. 15 |
| Building, construction, civil works | 8-12% | $60-95B | Ch. 16, 17 |
| Land, permitting, grid interconnection | 3-5% | $25-40B | Ch. 16 |
| Software, firmware, orchestration | <1% | <$8B | Ch. 5, 19 |
Note: Semiconductor equipment and materials do not receive hyperscaler capex directly; their revenue flows indirectly through wafer pricing at TSMC, Samsung, and Intel. The percentages above reflect the share of total AI infrastructure spending that ultimately reaches each layer, not direct hyperscaler procurement. Rows may not sum to 100% due to overlaps and exclusions (see table above for row definitions).
The key insight is that the GPU itself, while the highest-value single component, cascades demand through every other layer. A single NVIDIA GB200 NVL72 rack requires: 72 GPUs with HBM3e memory (SK Hynix, Samsung, Micron); advanced packaging (TSMC CoWoS) with ABF substrates (Unimicron, Ibiden); hundreds of optical transceivers at 800G or 1.6T (Coherent, Innolight); precision power delivery (Monolithic Power Systems VRMs, Infineon/ON Semi SiC); 120+ kW of cooling capacity per rack (Vertiv, CoolIT); dedicated electrical infrastructure (transformers, switchgear); and the physical building to house it all. This cascading demand structure is why the “picks and shovels” thesis extends far beyond the GPU makers themselves.
1.3 Power: The Binding Constraint
Power is the single most discussed bottleneck in the buildout, and the data supports the concern.
Current state: US data centers consumed approximately 176 TWh of electricity in 2023, representing about 4.4% of total US electricity consumption. This had tripled from 58 TWh in 2014 1112. The IEA estimates US data centers consumed 183 TWh in 2024, or slightly above 4% of total US electricity 13.
Projections: The DOE/LBNL report projects US data center electricity consumption will reach 325-580 TWh by 2028, or 6.7-12% of total US electricity 1112. The wide range reflects uncertainty about efficiency gains and buildout pace. EPRI estimates data centers could consume 4.6-9.1% of all US electricity by 2030 14. The IEA projects global data center electricity demand will approximately double by 2030, reaching roughly 945 TWh, equivalent to Japan’s entire current electricity consumption. In the US specifically, the IEA’s base case projects 426 TWh by 2030, a 133% increase from 2024 13. Goldman Sachs estimates a current 11 GW US power shortfall for data centers, widening to 40+ GW by 2028. McKinsey’s buildout analysis requires 125 GW of incremental AI capacity by 2030 15.
By 2030, powering data processing in the US is projected to consume more electricity than manufacturing all energy-intensive goods combined, including aluminum, steel, cement, and chemicals 16.
Geographic concentration amplifies the problem: In 2023, data centers consumed approximately 26% of Virginia’s total electricity supply, and significant shares in North Dakota (15%), Nebraska (12%), Iowa (11%), and Oregon (11%) 13. A voltage fluctuation in northern Virginia in July 2024 triggered the simultaneous disconnection of 60 data centers, causing a 1,500 MW power surplus that forced emergency grid adjustments to prevent cascading outages 17.
The transformer bottleneck within the power bottleneck: Average lead times for small-scale transformers reached approximately 120 weeks in 2024, with large power transformers taking as long as 210 weeks. Amazon has publicly noted transformer shortages delaying hyperscale data center builds in Virginia and Ohio 18. Oracle has also reportedly delayed some developments by as much as a year due to labor and material shortages 18.
Water is an under-discussed constraint: US data centers consumed approximately 17 billion gallons of water in 2023, primarily for evaporative cooling. Hyperscale and colocation facilities accounted for 84% of this. Hyperscale data centers alone are expected to consume 16-33 billion gallons annually by 2028 13.
The current energy mix for data centers: As of 2024, natural gas supplied over 40% of electricity for US data centers. Renewables (wind and solar) supplied about 24%, nuclear about 20%, and coal about 15%. Natural gas is projected to continue supplying the largest share through 2030, though nuclear could play a larger role if SMRs come online 13.
1.4 The Economic Logic
Why are hyperscalers willing to spend at these levels? The stated justification is that demand exceeds supply.
Google Cloud revenue grew 63% year-over-year to $20B in Q1 2026. Google’s cloud contract backlog reached $460 billion, roughly double the $240B at end of Q4 2025. CEO Sundar Pichai stated Google is “compute-constrained in the near term” 1. AWS revenue growth was 28% YoY in Q1 2026 4. All four hyperscalers reported demand exceeding available capacity in their most recent earnings calls.
At the broader level, McKinsey projects AI could create $2.6-4.4 trillion in annual economic value. Enterprise AI adoption has reached 87% among large organizations with 130% YoY spending growth 19. GitHub Copilot reached 1.8 million paying subscribers, with 46% of code now AI-generated 19.
The logic, stripped to first principles: if AI inference becomes cheap enough, it becomes cost-competitive with human labor for significant categories of service work. At that point, the total addressable market for AI compute is not “enterprise software budgets” but “labor markets,” which are orders of magnitude larger. Whether this logic holds is the central question of the entire buildout thesis.
1.5 The Bull Case
The strongest version of the bull case rests on four pillars.
Jevons Paradox (efficiency drives more consumption, not less): When AI per-unit costs fall, usage expands to fill the gap. Evidence from 2025: the cost to achieve a similar score on a challenging AI benchmark fell from $4,500 per task to $11.64 over the course of the year, yet total AI usage increased massively. A PIIE analysis found that “most firms allocate 90 percent of their total usage to a single model,” suggesting deep lock-in, and that AI usage has “dwarfed efficiency gains, just as Jevons Paradox would predict” 20.
Multimodal expansion: Text-to-text inference is the cheapest AI modality. Text-to-image, text-to-audio, and especially text-to-video require orders of magnitude more compute. As AI expands beyond chatbots into video generation, robotics, scientific simulation, and real-time decision-making, compute demand grows nonlinearly 21.
Infrastructure as competitive moat: The hyperscalers view this as a race they cannot afford to lose. Google co-founder Larry Page was quoted: “I’m willing to go bankrupt rather than lose this race” 22. Meta’s Zuckerberg, hours after DeepSeek’s release in January 2025, raised 2025 AI spending to $60-65B (+50% YoY), stating “scaling up infrastructure is still an important long-term advantage” 23.
Physical scarcity compounds: North American colocation vacancy reached 2.3% in H1 2025 (down from 9.8% in 2020) 15. Nearly 2 TW of clean energy (1.6 times current grid capacity) sits delayed in interconnection queues 16. Power, land, water, and skilled labor constraints mean that even if demand softens, new supply takes years to come online, creating a prolonged supply-demand imbalance.
1.6 The Bear Case
The strongest version of the bear case.
DeepSeek and the efficiency disruption: In January 2025, Chinese startup DeepSeek released R1, claiming training costs of roughly $6M versus roughly $600M for GPT-4, using export-restricted lower-tier NVIDIA chips. Performance appeared competitive with US frontier models. NVIDIA lost $589 billion in market cap in a single day, the largest single-day market cap loss in US stock market history. Broadcom, Micron, and Arm each fell more than 10%. Constellation Energy fell 20% 24.
The bear interpretation: if competitive models can be built at a fraction of the cost, the entire financial architecture of the buildout loses much of its justification. It rests on the assumption that the most powerful AI requires the most expensive hardware at ever-increasing scale. If that assumption breaks, stranded assets become a real risk 24.
The Jevons rebuttal (section 1.5) has been validated so far: hyperscaler capex accelerated after DeepSeek, not decelerated. But a critical qualification applies. The Jevons effect only holds if AI delivers genuine productivity gains that expand the market. If AI stalls at narrow incremental improvements that fail to translate into measurable economic output, efficiency stops compounding demand and overbuilt capacity chases diminishing marginal value 2125.
ROI uncertainty: Despite massive spending, the gap between capex and demonstrable AI revenue is widening. Capital intensity at 45-57% of revenue is historically unprecedented for technology companies 7. Free cash flow is collapsing at several hyperscalers (see section 1.1) 6.
Debt and financing risk: Hyperscalers raised $108B in debt in 2025. Projections suggest $1.5T in new debt issuance over coming years. Alphabet’s long-term debt quadrupled in 2025 to $46.5B 68. If AI revenue growth disappoints, the debt servicing burden could force spending pullbacks. Sustained high interest rates would compound this.
The replacement cycle overhang: AI accelerators have a competitive useful life of approximately three years. Physical failure is not the binding constraint; obsolescence is. An H100 deployed in 2023 is not broken in 2026, but it is uncompetitive against B200/B300-class hardware for frontier training workloads, and its inference cost per token is 3-5x worse than current-generation silicon. This creates a predictable depreciation cliff. The roughly $100B+ in H100-class hardware deployed during 2023-2024 reaches end-of-competitive-life by 2026-2027. By 2027-2028, hyperscalers face the simultaneous convergence of two capital demands: growth capex (expanding total installed compute to meet rising demand) and replacement capex (retiring and replacing hardware that has fallen below the competitive threshold). Under the bull case, where AI revenue grows fast enough to fund both streams, total capex plateaus at a high level but remains sustainable. Under the bear case, where AI revenue growth disappoints at precisely the moment the replacement cycle arrives, the dynamics become self-reinforcing in the wrong direction. If the “revenue wall” hits in 2027-2028, hyperscalers cannot defer replacement (the hardware is genuinely uncompetitive), cannot fund growth (cash flows are insufficient), and cannot maintain investor confidence in rising capex (the ROI narrative has broken). In this scenario, capex does not plateau; it collapses, because replacement spending without growth spending signals a business in runoff, not expansion. The companies most exposed to this timeline are the ODMs and OEMs (Quanta, Foxconn, Dell, Supermicro) whose revenue is directly and linearly tied to new hardware shipment volume. They have no recurring revenue stream to cushion a capex downcycle. Their stock prices reflect perpetual growth; a single year of flat-to-declining shipments would compress multiples severely. The scenario is conditional, not forecasted: it requires AI revenue growth to disappoint AND the replacement cycle to arrive simultaneously. But the timing alignment is plausible, and the market is not pricing it. Bubble historical parallels: Ray Dalio has called AI investment “very similar” to the dot-com bubble. The IMF has warned of an impending bust. Apollo’s chief economist notes that today’s top tech companies trade at higher valuations than their 1990s counterparts 24. The counterargument is that unlike the dot-com era, today’s AI companies generate real revenue and operate proven cloud businesses. The question is whether AI specifically, as opposed to cloud computing broadly, justifies the incremental spending.
Component cost inflation as a hidden tax: Microsoft attributed roughly $25B of its $190B 2026 capex to “higher component pricing,” primarily memory chips. Google and Meta also cited rising component costs in raising their spending plans 16. This means a portion of the capex increase reflects inflation, not incremental capacity.
1.7 Geopolitical Context
The buildout is happening within a geopolitically fractured environment that both accelerates and constrains it.
US CHIPS Act: The CHIPS and Science Act (August 2022) appropriated $52.7B for US semiconductor manufacturing and innovation, part of a larger roughly $280B legislative package 2627. Private companies have announced over $500B in investments to build, expand, or modernize chip facilities across 25+ US states 28. “Guardrails” restrict recipients from expanding semiconductor investments in “countries of concern” (China, Iran, Russia, North Korea) 26.
US export controls on China: Since October 2022, the US has progressively tightened controls on exports of advanced semiconductors, semiconductor manufacturing equipment, and EDA software to China 2729. Key milestones: October 2022 initial controls, October 2023 tightening, December 2024 further restrictions, March 2025 additional entity listings 29. The Netherlands restricted ASML from exporting advanced DUV lithography equipment to China (restrictions expanded in January 2025) 29. Japan added 23 items to its export control list 28.
In January 2026, the Trump administration shifted to case-by-case review (from presumption of denial) for H200-class and MI325X-equivalent chips to China, while simultaneously imposing 25% tariffs on imported semiconductors under Section 232 30. Eligible chips must have TPP below 21,000 and total DRAM bandwidth below 6,500 GB/s 30. This represents a partial relaxation of the chip export ban paired with tariff-based protection for domestic production.
China’s response: China has responded with massive state-backed investment ($150B+ channeled through the National Integrated Circuit Industry Investment Fund and other vehicles), restricted exports of gallium and germanium to the US (December 2024), and accelerated domestic chip development 2829. Huawei and SMIC partnered to produce a 7nm chip using equipment not restricted by export controls 29. SMIC has grown to become the world’s second-largest pure-play foundry by volume. China is expected to add more chipmaking capacity than the rest of the world combined in the coming years, though primarily at mature nodes (14nm and above) 28.
Implications for the buildout: The bifurcation creates two parallel technology ecosystems: a US-led sphere focused on the frontier, and a Chinese sphere driven by self-sufficiency 28. This does not reduce total global spending. If anything, duplication of supply chains increases total capex required. But TSMC in Taiwan remains the irreplaceable chokepoint for advanced node fabrication, creating a single geographic concentration risk that no amount of CHIPS Act funding has yet resolved.
1.8 Bottleneck Theory
Not all companies in the buildout capture value equally. The framework for identifying disproportionate pricing power is bottleneck analysis, rooted in Goldratt’s Theory of Constraints 31: system throughput is limited by its single tightest constraint, and improving non-constraint elements yields no benefit until the constraint is addressed. This report extends that framework with supply chain resilience theory (Christopher & Peck 32, Chopra & Sodhi 33) to map not just where constraints sit but how they interact, and applies FMEA methodology adapted from reliability engineering 37 to score their severity. The common-cause failure analysis draws on Lynch’s single-point-of-failure framework 36 and recent systematic reviews of cascading failures in supply chains 38. Two peer-reviewed reviews of semiconductor supply chain resilience specifically (Ramirez & Le 34, Azadegan et al. 35) confirm that the concentration risks identified in this report are consistent with the broader academic literature on semiconductor vulnerability.
Characteristics of a true bottleneck:
- Few suppliers (ideally 1-3) with high barriers to entry
- Long lead times that cannot be shortened with capital alone
- The bottleneck component is a small fraction of total system cost but is mission-critical (the buyer is price-insensitive)
- Switching costs are high (qualification processes take years)
- Capacity expansion takes longer than demand growth
Strong bottlenecks in this buildout (detailed analysis in Part II chapters):
- ASML: sole EUV lithography source; machines cost $200M+ 39; 18-24 month lead times [See Chapter 4]
- ABF substrates: five companies hold 74% of production; 30+ week lead times (see Chapter 9)
- EUV photoresists: Japanese producers hold roughly 91% global share; qualification takes years (see Chapter 3)
- Large power transformers: 120-210 week lead times; structural undersupply (see Chapter 14)
- HBM memory: three companies (SK Hynix, Samsung, Micron); SK Hynix has first-mover advantage (see Chapter 8)
- TSMC CoWoS packaging: demand exceeds capacity; NVIDIA allocation reportedly limited by CoWoS availability, not GPU die supply (see Chapter 9)
Weaker bottlenecks (important but more substitutable):
- Servers (multiple ODMs/OEMs), data center construction (many contractors, though skilled labor is scarce), fiber optic cable (Corning leads but capacity is less constrained), standard electrical components (commoditized).
The investment implication: within a $725B capex envelope 1, companies sitting at true bottlenecks capture disproportionate margin because buyers have no alternative but to pay. This framework guides the company analysis in Part II. For each theme chapter, I identify where the bottleneck sits and which companies occupy it.
Key Numbers Summary
| Metric | Figure | Source |
|---|---|---|
| Big Four 2026 capex (AMZN/GOOG/MSFT/META) | ~$725B | 1 |
| Top 9 CSP 2026 capex (incl. Oracle, Chinese) | ~$830B | 2 |
| AI share of hyperscaler capex | ~75% | 7 |
| Cumulative 2025-2027 hyperscaler capex | ~$1.15T | 8 |
| US DC electricity 2023 | 176 TWh (4.4% of US) | 1112 |
| US DC electricity 2028 (projected) | 325-580 TWh (6.7-12%) | 1112 |
| Global DC electricity 2030 (IEA base) | ~945 TWh | 13 |
| US DC water consumption 2023 | ~17B gallons | 13 |
| Large power transformer lead time | Up to 210 weeks | 18 |
| NA colocation vacancy (H1 2025) | 2.3% | 15 |
| DeepSeek R1 training cost claim | ~$6M | 24 |
| NVIDIA single-day loss (Jan 27, 2025) | $589B | 24 |
| CHIPS Act appropriation | $52.7B | 2627 |