Jensen Huang’s CES 2026 framing shifted the AI economics conversation: AI factories don’t produce tokens—they produce actions.
Traditional AI Economics (LLMs)
| Factor | Value |
|---|---|
| Core Metric | Cost per token |
| Workload Pattern | Inference is lightweight vs. training |
| Scaling Formula | Model parameters x Tokens generated |
| Inference Mode | Batch, asynchronous (100ms+ acceptable) |
Physical AI Economics
| Factor | Value |
|---|---|
| Core Metric | Cost per successful action |
| Workload Pattern | Training, simulation, AND inference run continuously |
| Scaling Formula | (Parameters x Simulation fidelity x Action frequency) x Deployed robots |
| Inference Mode | Real-time, continuous, safety-critical (sub-ms required) |
The Infrastructure Cascade
- Edge chips become critical path: Jetson Thor (1 PFLOP) enables on-device humanoid reasoning. 542,000 robots x continuous inference = massive aggregate edge demand.
- Connectivity becomes non-negotiable: NVIDIA’s partnership with T-Mobile, Cisco for 6G. Robot fleets need industrial-grade networking.
- Data center topology shifts: Physical AI needs training + simulation + edge triangulation, not just centralized inference.
The Flywheel
Physical AI isn’t just a market for AI compute—it’s becoming the primary generator of grounded, physics-based training data that LLMs and agents need to move beyond pure language reasoning.
This analysis is part of a comprehensive report. Read the full analysis: Physical AI Is Crossing the Manufacturing Chasm on The Business Engineer.









