Microsoft Maya 200
by Brian Wang from NextBigFuture.com on (#735A8)
, the Maia 200 packs 140+ billion transistors, 216 GB of HBM3E, and a massive 272 MB of on-chip SRAM to tackle the efficiency crisis in real-time inference. Hyperscalers prioritize inference efficiency and cost (40-50% reductions). By 2028, custom ASICs could capture 20-30% market from Nvidia's ~90%, with total AI chip sales ~$975B in 2026. ...