Nvidia released benchmark data showing its latest AI server, which packs 72 of its top chips into a single system, can deliver roughly a 10x performance gain when serving large mixture‑of‑experts models such as Moonshot AI’s Kimi K2 Thinking and DeepSeek’s models. The results aim to show that even as some new models train more efficiently, Nvidia’s high‑end servers remain critical for large‑scale inference, reinforcing its dominance against rivals like AMD and Cerebras in the AI deployment market.
This article aggregates reporting from 1 news source. The TL;DR is AI-generated from original reporting. Race to AGI's analysis provides editorial context on implications for AGI development.


