Maia 200 AI chip promises threefold FP4 power, edging out TPU and Trainium in inference

1 min read
Source: Live Science
Maia 200 AI chip promises threefold FP4 power, edging out TPU and Trainium in inference
Photo: Live Science
TL;DR Summary

Microsoft unveiled Maia 200, an AI inference accelerator for Azure, claiming it delivers over 10 petaflops at FP4 and 5 PFLOPS at FP8, with 3x FP4 performance versus Amazon’s Trainium Gen3 and FP8 performance above Google’s TPU Gen7. Built on TSMC’s 3-nanometer process with about 100 billion transistors, Maia 200 is designed for data-center inference to speed Copilot and Azure OpenAI workloads, featuring a memory system to keep model weights local and is currently deployed in a US data center with broader Azure availability planned in the future.

Share this article

Reading Insights

Total Reads

0

Unique Readers

15

Time Saved

55 min

vs 56 min read

Condensed

99%

11,15288 words

Want the full story? Read the original article

Read on Live Science