Maia 200: Microsoft Built Its Own Chip Because Inference Costs Were the Problem
Every conversation about enterprise AI eventually hits the same wall: "What does this cost at scale?"
Training a model is a one-off expense. Inference — actually running that model in production, millions of times a day — is the ongoing bill that kills business cases. And for most organisations evaluating AI, the inference economics don't yet work.
Microsoft just threw 140 billion transistors at that problem.
What Maia 200 actually is
Maia 200 is Microsoft's custom-designed AI inference accelerator. Not a GPU. Not a general-purpose chip. A silicon design built from scratch, specifically to generate tokens cheaper and faster than anything else in Azure's fleet.
The numbers are significant. Built on TSMC's 3nm process, each chip delivers over 10 petaFLOPS in FP4 precision and over 5 petaFLOPS in FP8, within a 750W power envelope. It packs 216GB of HBM3e memory at 7 TB/s bandwidth, plus 272MB of on-chip SRAM to keep data close to the compute.
Microsoft claims 30% better performance per dollar than the latest-generation hardware in their existing fleet. They also claim 3x the FP4 performance of Amazon's third-generation Trainium and FP8 performance above Google's seventh-generation TPU.
Those are bold comparisons. But the architecture backs them up.
Why custom silicon matters now
The AI industry has operated on a simple model: rent NVIDIA GPUs, run your workloads, pay the bill. That worked when AI was experimental. It doesn't work when you're running GPT-5.2 inference across Microsoft 365 Copilot for millions of users.
Microsoft's approach with Maia 200 is vertical integration. Design the chip, the networking, the cooling, the software stack, and the datacentre integration as one system. The interconnect is standard Ethernet — not a proprietary fabric — which keeps switching costs down and enables commodity networking equipment. Each accelerator exposes 2.8 TB/s of bidirectional scale-up bandwidth, supporting clusters of up to 6,144 accelerators.
The architecture is deliberately optimised for narrow-precision datatypes. FP4 and FP8 inference maintains model accuracy while cutting compute and memory requirements significantly. Hardware-based data casting converts storage types to compute types at line rate, so there's no performance penalty for storing tensors in lower precision.
This is the kind of optimisation that only makes sense when you control the full stack. You can't do this with off-the-shelf GPUs.
The heterogeneous fleet strategy
Maia 200 doesn't replace NVIDIA GPUs. It sits alongside them. Microsoft is building a heterogeneous AI infrastructure — Maia for inference, NVIDIA for training, AMD for specific workloads — with the Azure control plane routing work to the most cost-effective hardware automatically.
This is similar to what the Model Router does for models in Foundry (routing prompts to the cheapest capable model), but at the infrastructure layer. The right silicon for the right workload, managed transparently.
For enterprise customers, this means AI inference costs should come down without any changes to your code. The Foundry API stays the same. The model stays the same. The hardware underneath gets cheaper. That's the promise, at least.
The Maia SDK
Microsoft is previewing the Maia SDK for developers, AI startups, and academics who want to optimise directly for the hardware. It includes PyTorch integration, a Triton compiler, optimised kernel libraries, and access to Maia's low-level programming language (NPL). There's also a Maia simulator and cost calculator for estimating efficiency gains before committing.
This is an interesting move. Most cloud customers won't touch hardware-level optimisation. But for teams building high-volume inference pipelines — synthetic data generation, real-time recommendation engines, large-scale document processing — the SDK opens a path to squeezing out additional performance. Microsoft's own Superintelligence team is already using Maia 200 for synthetic data generation and reinforcement learning.
Sign up for the SDK preview at aka.ms/Maia200SDK.
Getting started
Maia 200 is deployed in Azure's US Central region (Des Moines, Iowa), with US West 3 (Phoenix, Arizona) coming next. Global expansion is planned but no timeline has been shared.
For most enterprise customers, the immediate impact is indirect. You won't select "Maia 200" in the Azure portal. You'll deploy a model in Foundry, and Azure will route your inference to the most efficient hardware available — which increasingly includes Maia 200. The cost benefits flow through to your existing AI workloads.
If you want direct access to optimise for the hardware, the SDK preview is the path. It's early — preview-stage tooling with the usual caveats — but it's a real opportunity for teams with high-volume inference requirements.
What this means for AI economics
The AI infrastructure competition has entered its custom-silicon phase. Google has TPUs. Amazon has Trainium and Inferentia. Microsoft now has Maia. Each hyperscaler is betting that controlling the silicon is the key to controlling AI costs.
For IT leaders, the practical takeaway is straightforward: inference costs are coming down. Not incrementally — structurally. The 30% performance-per-dollar improvement Microsoft claims with Maia 200 is a first generation. They've confirmed this is a multi-generational programme with future chips already in design.
If your AI business case was marginal because of inference costs, it's worth revisiting the numbers. The infrastructure economics are moving in your favour. And unlike model improvements — which change what's possible — silicon improvements change what's affordable.
That might be the more important shift.
Leon Godwin is Principal Cloud Evangelist at Cloud Direct, helping organisations build cloud strategy with clarity and technical honesty.