This analysis is part of Google’s AI Full-Stack Domination, a deep dive by The Business Engineer.

Everything in Google’s AI stack starts with chips. Google designed TPUs specifically for its own AI workloads, giving it a 2-3 year design iteration advantage over companies dependent on NVIDIA’s allocation cycles.
TPU v7 Ironwood
Ironwood is approaching general availability. Validation is extraordinary: Anthropic is committed to access up to 1 million TPUs, and 9 of the top 10 AI labs now choose Google Cloud infrastructure. This is not a commodity chip business—it’s a vertically integrated compute layer where hardware is co-optimized with JAX and TensorFlow.
The 78% Number
Gemini serving costs dropped 78% in 2025, driven by model optimizations running on proprietary hardware. That single number explains more about Alphabet’s strategic position than any revenue figure. When your foundation gets 78% cheaper in one year while competitors pay market rates for GPU access, every product built on top carries a structural margin advantage.
The Compounding Loop
Invest in custom silicon → 78% cost reduction compounds → every product gets cheaper → more revenue → more chips → repeat. CapEx going from $91B (2025) to $175B+ (2026) isn’t excessive spending. It’s scaling the foundation that makes everything above it cheaper.
The silicon layer creates the cost gravity that pulls margins upward across the entire stack.








