This post is also available in:
Malayalam
Microsoft has announced its newest in-house silicon, the Maia 200, describing it as a powerful processor designed to scale AI inference workloads more efficiently.
The Maia 200 succeeds the Maia 100, which debuted in 2023, and is engineered to run advanced AI models at higher speeds while consuming less power. According to Microsoft, the new chip contains more than 100 billion transistors and delivers over 10 petaflops of performance at 4-bit precision, along with roughly 5 petaflops at 8-bit precision—representing a major performance jump over the previous generation.
Inference refers to the process of running trained AI models, as opposed to training them. As AI companies grow and deploy models at scale, inference has become a significant contributor to overall computing costs, driving demand for more efficient hardware solutions.
Microsoft believes the Maia 200 can help address these challenges by enabling AI systems to operate more smoothly while reducing power consumption. The company said that a single Maia 200 node is capable of running today’s largest AI models comfortably, with capacity to support even more complex models in the future.
The launch of Maia 200 also reflects a broader industry shift, with major technology firms increasingly developing custom chips to reduce reliance on Nvidia’s GPUs, which dominate the AI hardware market. Google offers its Tensor Processing Units (TPUs) through cloud-based compute services, while Amazon has developed its own AI accelerators under the Trainium brand, most recently introducing Trainium3 in December. These custom chips help offload workloads that would otherwise require Nvidia hardware, lowering costs.
With Maia, Microsoft aims to compete directly with these alternatives. In its announcement, the company claimed that Maia 200 delivers three times the FP4 performance of Amazon’s third-generation Trainium chips and exceeds the FP8 performance of Google’s seventh-generation TPU.
Microsoft added that the Maia chip is already being used to power AI models developed by its Superintelligence team and to support Copilot, the company’s AI assistant. The company has also begun inviting developers, researchers, academic institutions, and leading AI labs to use the Maia 200 software development kit for their own workloads.
This post is also available in:
Malayalam

