
The Training-to-Inference Inversion
The AI market’s center of gravity is shifting. In 2023, training represented two-thirds of compute workloads. In 2026, inference does. The inference-optimized chip market is projected to exceed $50B in 2026 and $167B by 2032 (28.25% CAGR). This is not a cyclical rotation — it is a structural inversion of where value concentrates.
In a training-dominated market, the competitive axis is raw FLOPS and the software ecosystem (CUDA). In an inference-dominated market, the competitive axes shift to cost-per-token, power efficiency, latency, and deployment flexibility. These are precisely Qualcomm’s core competencies — refined over two decades of mobile silicon engineering.
Amon framed it through the economics of necessity at Davos: “If you’re a company spending billions of dollars building a data center for training, you expect to get a return on that investment. So when you start putting AI into production, you’re doing a lot of inference. The inference is a significantly growing opportunity.”
Jack Gold of J Gold Associates predicts that within two to three years, 85% of enterprise AI workloads will be inference-based. If that proves correct, Nvidia’s 92% data center share — earned primarily through training dominance — becomes increasingly misaligned with where the majority of compute demand sits.
What Changes When Inference Dominates
The shift from training to inference changes the competitive dynamics fundamentally:
- Training rewards concentration — massive GPU clusters, centralized data centers, CUDA lock-in
- Inference rewards distribution — diverse deployment environments, power efficiency, cost optimization
Amon drew an analogy to smartphone engineering at Davos: in a phone, you can’t run everything on the CPU because it burns too much power. So Qualcomm perfected “heterogeneous compute” — dedicated engines for music decode, video processing, camera ISP, each optimized for its specific workload. The same principle applies to inference.
There’s an architecture for prefill, another for decode. The GPU-centric monolith is giving way to specialized engines for specialized tasks. “We’re building what we believe is post-GPU,” Amon said.
The bet: cost-per-inference and power efficiency will matter more than raw throughput as inference workloads scale. The market currently prices this at near-zero revenue for Qualcomm in data center inference; analysts see $10B+ potential if executed.
The Memory Paradox
Q2 FY26 guidance came in below expectations because AI data center demand for HBM memory is creating an industry-wide shortage that constrains smartphone supply chains. Amon said it clearly: “Memory is going to define the size of the handset market.”
This is the AI market contradicting itself. Training and inference are now so compute-intensive that they starve adjacent markets of critical components. The very force pressuring Qualcomm’s near-term handset business confirms the long-term inference opportunity.
Training and inference now create structural tension. The transition is not theoretical — it is happening in real time, visible in supply chain constraints.
This is part of a comprehensive analysis. Read the full analysis on The Business Engineer.









