Research Question

Compile Q4 2025 and Q1 2026 CapEx guidance from major hyperscalers (Microsoft, Google, Amazon, Meta) and analyze any revisions or commentary about AI infrastructure spending. Distinguish between AI training vs. inference spending trends and assess whether guidance signals continued acceleration or early signs of moderation. Summarize implications for DRAM demand.

Microsoft Azure CapEx Acceleration: Reversal from Moderation Signals Training Buildout Priority

Microsoft reversed prior expectations of CapEx moderation by reporting record quarterly spending—$35B in FY26 Q1 (Jul-Sep 2025) and $37.5B in Q2 (Oct-Dec 2025)—with two-thirds on short-lived GPU/CPU assets for AI, driven by demand exceeding supply in a "planet-scale cloud and AI factory" that links data centers via AI-optimized WANs for massive training clusters; this front-loaded mechanism prioritizes training capacity now (adding ~1GW in Q2 alone) to unlock inference scale later, but capacity constraints cap Azure growth at 37-40% despite stronger underlying demand.[1][2][3]
- FY25 total AI/data center CapEx ~$80B; FY26 run-rate implies $140-150B+ (Q1-Q2 already $72.4B), up ~80% YoY[4][5]
- CFO Amy Hood: Growth "could have been higher without capacity constraints" persisting through FY26 (Jun 2026); Q3 CapEx to dip seasonally but overall rising[1]
- No explicit training/inference split, but emphasis on GPUs/CPUs and R&D allocation points to training dominance amid backlog doubling to $625B[2]

Implications for competitors/entrants: New players can't match Microsoft's data moat (real-time merchant/OpenAI integration) without years of buildout; focus on inference niches (e.g., edge AI) where training lock-in is weaker, but expect margin pressure from GPU scarcity.

Amazon AWS CapEx Surge: Inference Economics Justify $200B Bet on Custom Silicon

Amazon's Q4 2025 (calendar Q4) CapEx hit $40.5B (full 2025 ~$131-135B), guiding $200B for 2026—predominantly AWS—via Trainium/Inferentia chips that slash inference costs (structured silicon/depreciation/power optimization), monetizing capacity instantly against a $244B backlog up 40% YoY; this creates a flywheel where AI workloads (core + generative) drive 24% AWS growth at 35% margins, with inference poised as "majority of long-term AI workloads."[6][7][8]
- Added 3.9GW power in past 12 months (double 2022 levels), plans to double again by 2027; Q1 2026 revenue guide $173.5-178.5B implies sustained acceleration[9][10]
- CEO Andy Jassy: "Very high demand...monetizing as fast as we install it," with Trainium for training/inference reducing Nvidia reliance[6]

Implications for competitors/entrants: AWS's inference cost edge (via custom chips) pressures pure-play GPU resellers; entrants should target hybrid cloud niches but brace for 50%+ CapEx YoY forcing debt reliance.

Alphabet (Google Cloud) Doubles Down: 60/40 Server/Data Center Split Fuels Frontier Models

Alphabet's Q4 2025 CapEx was $27.9B (full 2025 $91.4B), guiding $175-185B for 2026 (~double YoY) with 60% servers (AI compute for DeepMind frontier training) and 40% data centers/networking, addressing a $240B Cloud backlog amid 48% GCP growth from AI infra/solutions; supply constraints persist through 2026 despite ramps, signaling no moderation.[11][12][13]
- Gemini processes 10B+ tokens/min via API; Cloud at $70B run-rate, tight supply caps faster growth[12]
- CEO Sundar Pichai: "Investing in AI compute...to meet customer demand," with rigorous ROI process[13]

Implications for competitors/entrants: Google's integrated stack (TPUs + Gemini) raises training barriers; specialize in inference optimization or regional clouds to avoid head-on compute wars.

Meta Platforms' Superintelligence Push: MTIA Extends to Training Amid 73% CapEx Jump

Meta's Q4 2025 CapEx reached $22.1B (full 2025 $72.2B), guiding $115-135B for 2026 (60%+ YoY) for Meta Superintelligence Labs via flexible infra (Nvidia/AMD/MTIA chips), where MTIA shifts from inference to training workloads in Q1 2026 while Andromeda triples efficiency; ad revenue funds this without FCF strain, targeting op income >2025.[14][15]
- Infrastructure costs (cloud/depreciation/OPEX) drive expense growth; Q1 2026 revenue $53.5-56.5B[15]

Implications for competitors/entrants: Meta's silicon/energy optimization lowers cost/GW; open-source Llama lowers entry for inference apps, but training scale requires hyperscaler alliances.

Hyperscalers' combined 2026 CapEx hits $635-700B (60-74% YoY from ~$380B 2025), up from prior moderation signals (e.g., Microsoft's FY26 reversal), with no Q4 2025/Q1 2026 quarterly guides but sequential Q3 dips noted; demand backlogs ($1.6T+ total) and capacity constraints confirm acceleration, training dominant now (GPUs, frontier models) but inference "majority long-term" per Amazon/Meta.[5][16]
- Revisions upward across board; no moderation signs despite investor FCF fears[17]

Implications for competitors/entrants: Acceleration validates AI supercycle; partner with leaders or niche in inference to sidestep training arms race.

DRAM Demand Boom: HBM Shift Starves DDR, Prices Triple into 2026 Shortage

Hyperscalers' GPU-heavy CapEx (60%+ servers) devours HBM/DDR5 (3x wafer for HBM vs DDR), reallocating 20-70% global capacity to AI servers and causing DDR shortages/inventories to 8 weeks; prices tripled YoY (DDR5 chips $6.84→$27+), persisting 2026-28 as fabs lag (Micron Japan 2028).[18][19][20]
- AI takes 20% wafers 2026; supply +16% vs demand, HBM shortfall widens to 9% by 2027[19]

Implications for competitors/entrants: Stockpile DDR/HBM now (prices +47% 2026); inference apps less memory-intensive offer breathing room vs training clusters. Confidence: High on CapEx (direct guidance), medium on splits (inferred from commentary), web-verified data current as Feb 2026.


Recent Findings Supplement (February 2026)

Hyperscaler CapEx Surge Signals Unabated AI Acceleration into 2026

Major hyperscalers—Microsoft, Alphabet (Google), Amazon (AWS), and Meta—collectively spent over $233 billion on CapEx in calendar 2025 (Q4 data finalized Jan-Feb 2026), up sharply from prior years, and guided for $650-700 billion in 2026, a 60%+ increase driven by AI compute demand outpacing supply. This front-loaded spend on GPUs/CPUs (short-lived assets ~2/3 of recent quarters) reflects a mechanism where hyperscalers are building "AI factories" via massive server deployments (e.g., Google's 60% servers/40% data centers split) to capture training and inference workloads before competitors, despite stock selloffs on near-term margin pressure; non-obvious implication: power constraints (e.g., Microsoft's $80B Azure backlog unfilled) force even higher future CapEx as capacity utilization lags demand.[1][2]
- Meta Q4 2025: $22.1B CapEx (FY2025 total $72.2B); 2026 guide $115-135B (+60-87% YoY), tied to Meta Superintelligence Labs and core AI infra.[3]
- Microsoft Q2 FY2026 (Q4 cal. 2025): $37.5B CapEx (66% YoY rise, 2/3 GPUs/CPUs); FY2026 run-rate ~$140-150B implied (no formal guide, but sequential Q3 decrease expected on timing).[4][5]
- Alphabet Q4 2025: $27.9B CapEx (FY2025 $91.4B); 2026 guide $175-185B (nearly double), ramping quarterly for AI compute.[2]
- Amazon Q4 2025: $39.5B CapEx (FY2025 $131.8B); 2026 guide $200B (mostly AWS), with Trainium chips powering inference/training (e.g., Bedrock majority inference).[6]
Implications for competitors/entrants: New players face insurmountable data moats; only suppliers (Nvidia, memory makers) benefit short-term, but hyperscalers' custom silicon (Amazon Trainium3/4, Meta MTIA) erodes GPU reliance by 2027, pressuring pure-play AI chipmakers.

No Moderation: Q4 2025 Marks Peak Acceleration, Q1 2026 Guides Confirm

Q4 2025 CapEx hit records across the board (e.g., Microsoft's $37.5B quarterly high), with no downward revisions—instead, upward surprises vs. consensus (Amazon $200B vs. $146B expected)—signaling demand exceeds even aggressive builds; mechanism: hyperscalers auto-allocate incoming GPUs to Azure/AWS capacity (customer demand > supply), replacing EOL gear while expanding R&D/first-party AI (Copilot), creating a self-reinforcing cycle where backlog growth (Amazon $244B AWS, +40% YoY) justifies more spend. Unlike 2023-2024 training ramps, 2026 commentary emphasizes monetization (e.g., Meta op. income >2025 despite CapEx step-up), implying sustained multi-year trajectory without peak signals.[7][8]
- All firms raised 2025 guides multiple times (Alphabet 3x to $91B); 2026 formalizes ~60% aggregate jump, with Q1 ramps (Alphabet depreciation accelerates Q1).
- No cuts: Sequential Q3 FY26 dip for Microsoft is timing-only (finance leases), not demand fade.[5]
Implications: Entrants must partner (e.g., Oracle $50B CapEx via hyperscaler deals) or niche (edge inference); overbuilders risk stranded assets if power/GPU delays persist into 2027.

Training Dominates 2025-26 Buildout, Inference Emerges as Durable Tailwind

Hyperscalers' 2025-26 CapEx remains ~training-heavy (frontier models like Google's DeepMind, Meta GEM scaling, Amazon Project Rainier/Claude), with ~60-67% on servers/GPUs for massive clusters (Microsoft 1GW "super factory" Q2 FY26 alone); however, inference trends signal shift: Amazon Trainium2/3 powers Bedrock majority inference (1.4M chips subscribed), Meta MTIA extends to training Q1 2026 but already runs ad ranking inference, implying inference's recurring nature (daily queries vs. one-time training) sustains post-2026 spend at 70-80% of peak. No explicit splits, but commentary ties CapEx to "core + AI workloads" growing faster than anticipated.[9][6]
- Training: Google 2026 CapEx supports DeepMind frontiers; Amazon doubles capacity to 2027.
- Inference: AWS "majority long-term AI workloads"; Meta ads quality +12% via inference.
Implications: Inference focus favors efficient custom chips (Trainium cost 30-50% below Nvidia), commoditizing suppliers; competitors need hybrid GPU/in-house stacks to match utilization.

DRAM Demand Explodes on AI Server Shortages, No Relief Until 2027

Hyperscalers' GPU/server frenzy diverts memory fab capacity to HBM (all 2026 sold out), spiking server DRAM prices 55-60% Q4 2025-Q1 2026 (TrendForce), with DDR5/LPDDR5X up 90-100% on AI server needs (e.g., Nvidia shift doubles server memory); mechanism: HBM prioritization starves commodity DRAM (Samsung/SK Hynix capex ~$54B all AI-focused), tightening supply to 8 weeks as hyperscalers stockpile for clusters—non-obvious: this "memory tax" adds 5-10% to cloud costs Q2-Q3 2026, indirectly boosting hyperscaler pricing power. Capacity online late 2027 earliest.[10][11]
- Server DRAM: +60-70% Q1 2026 contracts; inventories critical.
- HBM/DRAM bit demand: +35% 2026 vs. 23% supply (AI servers).
Implications: DRAM makers (Micron/SK Hynix) see margins >60%; entrants face 40-50% BOM hikes, widening hyperscaler moat—compete via software efficiency or wait for 2028 oversupply.

Aggregate 2026 Outlook: $700B Bet Validates AI as Core, Not Speculative

Combined hyperscaler CapEx (~$700B, +60% YoY) exceeds prior tech cycles (0.8-1.5% GDP), funded by cash flows/leases despite FCF squeezes, with ROIC assured via immediate Azure/AWS monetization (e.g., Amazon "as fast as installed"); cause-effect: AI diffusion (Gemini 10B tokens/min, Copilot backlog) drives cloud re-accel (AWS 24% Q4), forcing CapEx to close supply gaps—differing now: inference durability + custom silicon reduce Nvidia dependency, extending cycle vs. 2023 hype. Confidence high on guidance; risks in power/execution noted.[7]
- Breakdown: Amazon $200B (AWS lead), Alphabet $180B mid, Meta $125B mid, Microsoft ~$145B run-rate.[12]
Implications: Suppliers thrive 2026 (DRAM upcycle), but pure AI plays vulnerable post-inference shift; entrants: focus sovereign clouds or inference niches for scraps. Additional research: Q1 2026 earnings for execution proof.