Amazon prioritizes cost reduction for massive internal AI workloads. $125B CapEx signals long-term custom silicon commitment.
Trainium — AI Training Chip
- Deployment: 2.5 Million Chips projected 2024-2025
- Purpose: Large Model Training
- Cost/H100e: $7,200
Inferentia — AI Inference Chip
- Optimized for: Production Inference
- Use Cases: Low-latency, high-throughput (Alexa, Prime Video, Search)
Key Metrics
- Compute Share: 10.9%
- Revenue Share: 3.2%
- Revenue: $9.7B
Low revenue share relative to compute share = aggressive internal cost optimization
2025 Capital Expenditure
- Total CapEx: $125B
- YoY Increase: +60%
- Rank: #1 Largest hyperscaler spender
Internal Use Cases
- Amazon.com: Product recommendations, search, personalization
- AWS Services: Bedrock, SageMaker, customer AI workloads
- Alexa & Devices: Voice AI, smart home, edge inference
The Amazon Strategy
- Internal First: Optimize AWS costs before external sales
- Reduce NVIDIA Dependency: Control costs and supply chain
- Massive Scale Advantage: 2.5M chips = internal optimization at scale
Anthropic Partnership
- Investment: $8B+
- Integration: Claude on AWS
- Optimization: Trainium-tuned
Limitations
- Internal focus limits scale
- Less mature than TPU
- Still NVIDIA-dependent for some workloads
Strategic Position: Internal Optimization Leader. Amazon prioritizes cost reduction for massive internal AI workloads.
This is part of a comprehensive analysis. Read the full analysis on The Business Engineer.








