custom-silicon

2 items

GeekWire 2026-03-23-3

AWS at 20: Inside the rise of Amazon's cloud empire, and what's at stake in the AI era

GeekWire's oral history buries the competitive signal inside the nostalgia: AWS customers are bypassing Bedrock to call Anthropic directly, which means the fastest-growing AWS service ever may be growing on committed-spend burn-down, not organic AI workload choice. The $200B capex bet and Jassy's $600B revenue target are Amazon paying to stay relevant at a stack layer it used to own; the structural question is whether AWS becomes a platform or a utility as models become the new developer interface. Azure at $75B (34% growth), Google Cloud at $50B, and the OpenAI deal at 16x Microsoft's per-point cost all point the same direction: the cloud market AWS created is converging, and custom silicon is the last defensible layer.

New York Times 2026-03-17-3

Nvidia Built the A.I. Era. Now It Has to Defend It.

Nvidia is the first major chipmaker to unbundle training from inference at the architecture level, pairing its GPUs with Groq's inference-optimized LPUs in a $20B licensing deal. The supply chain math is as interesting as the product: Groq on Samsung fab with no HBM dependency sidesteps both TSMC allocation constraints and memory chip shortages. If inference grows to 70-80% of total AI compute spend, the companies building chip-agnostic inference routing will capture a new middleware layer that doesn't exist yet.