Although Nvidia has dominated the AI chip market for years, Microsoft’s announcement of the Maia 200 on January 26, 2026, signals a new challenge that could shake up the playing field. Yet Nvidia’s stock price barely budged. Why did investors stay so calm when a tech giant disclosed such powerful hardware?
The answer lies in what the Maia 200 actually does. Microsoft built this chip specifically for AI inference, not training. Inference means running AI models that already exist, like when someone asks a chatbot a question. Training means building new AI models from scratch, which requires different hardware. Nvidia still rules the training world, where companies spend the most money developing cutting-edge AI systems. This distinction is reinforced by the way futures and derivatives markets help firms hedge large capital expenditures with standardized contracts.
The Maia 200 packs impressive specs. It has over 100 billion transistors built on TSMC’s 3nm process and delivers more than 10 petaFLOPS in 4-bit precision. That’s three times faster than Amazon’s Trainium chip for certain tasks. With 216GB of HBM3e memory running at 7TB/s, it handles massive AI models efficiently. Microsoft designed it to power services like Copilot and GPT-5.2 in Azure datacenters, starting in the US Central region.
However, investors understand that custom inference chips serve a specific purpose. Companies like Microsoft, Google, and Amazon build these chips to run their own services more cheaply. They reduce costs for operations they control but don’t necessarily replace Nvidia’s broader ecosystem. Nvidia sells chips to thousands of companies worldwide, not just a handful of tech giants.
Additionally, the Maia 200 works best with low-precision calculations, which save power but limit flexibility. Nvidia’s GPUs handle both training and inference across many precision levels, making them more versatile. Think of it like owning a specialized tool versus a Swiss Army knife. Both have value, but one serves more situations.
Microsoft’s chip represents progress in silicon competition, but it targets a narrower market segment. Nvidia investors recognized that inference-only chips don’t threaten the core business of selling general-purpose AI accelerators. Microsoft has already invited developers, academics, and frontier AI labs to use the Maia 200 software development kit for external workloads. Microsoft confirmed it will continue to buy Nvidia and AMD chips despite having the Maia 200, signaling ongoing reliance on external suppliers. That’s why the stock stayed steady despite Microsoft’s technological achievement.




