Top 10 most intelligent open-source Artificial Intelligence models run 10x faster on NVIDIA Blackwell NVL72

Mixture-of-experts architectures power the top 10 open-source Artificial Intelligence models on the Artificial Analysis (AA) leaderboard. NVIDIA says its GB200 NVL72 rack-scale system delivers a 10x performance leap for the Kimi K2 Thinking model versus HGX H200.

A look under the hood of many frontier models shows a preference for mixture-of-experts architectures. The article explains that mixture-of-experts models mimic the human brain by activating specific experts for each token, which reduces compute requirements while increasing token-generation efficiency. On the independent Artificial Analysis (AA) leaderboard, the top 10 most intelligent open-source models use an MoE architecture, and the article lists DeepSeek Artificial Intelligence’s DeepSeek-R1, Moonshot Artificial Intelligence’s Kimi K2 Thinking, OpenArtificial Intelligence’s gpt-oss-120B and Mistral Artificial Intelligence’s Mistral Large 3 as examples among that top group.

Scaling mixture-of-experts models in production is described as notoriously difficult because achieving both high efficiency and high performance requires close coordination between hardware and software. The article highlights NVIDIA’s GB200 NVL72 rack-scale systems as an extreme codesign that combines hardware and software optimizations to make MoE scaling practical. It reports a specific performance comparison: the Kimi K2 Thinking MoE model, ranked as the most intelligent open-source model on the AA leaderboard, sees a 10x performance leap on the NVIDIA GB200 NVL72 rack-scale system compared with NVIDIA HGX H200. The piece frames that gain as a demonstration of how system-level engineering can unlock the efficiency benefits of MoE architectures.

Finally, the article ties the NVL72 results to other MoE deployments, saying the breakthrough builds on performance delivered for DeepSeek-R1 and Mistral Large 3 MoE models. It concludes that mixture-of-experts is becoming the architecture of choice for frontier models and positions NVIDIA’s full-stack inference platform as key to realizing the architecture’s potential in production environments.

63

Impact Score

Samsung’s 2 nm node progress could revive foundry business and attract Qualcomm

Samsung Foundry’s 2 nm SF2 process is reportedly stabilizing at around 50% yields, positioning the Exynos 2600 as a key proof of concept and potentially helping the chip division return to profit. New demand from Tesla Artificial Intelligence chips and possible deals with Qualcomm and AMD are seen as central to the turnaround.

How high quality sound shapes virtual communication and trust

As virtual meetings, classes, and content become routine, researchers and audio leaders argue that sound quality is now central to how we judge credibility, intelligence, and trust. Advances in Artificial Intelligence powered audio processing are making clear, unobtrusive sound both more critical and more accessible across work, education, and marketing.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.