Top 10 most intelligent open-source Artificial Intelligence models run 10x faster on NVIDIA Blackwell NVL72

Mixture-of-experts architectures power the top 10 open-source Artificial Intelligence models on the Artificial Analysis (AA) leaderboard. NVIDIA says its GB200 NVL72 rack-scale system delivers a 10x performance leap for the Kimi K2 Thinking model versus HGX H200.

A look under the hood of many frontier models shows a preference for mixture-of-experts architectures. The article explains that mixture-of-experts models mimic the human brain by activating specific experts for each token, which reduces compute requirements while increasing token-generation efficiency. On the independent Artificial Analysis (AA) leaderboard, the top 10 most intelligent open-source models use an MoE architecture, and the article lists DeepSeek Artificial Intelligence’s DeepSeek-R1, Moonshot Artificial Intelligence’s Kimi K2 Thinking, OpenArtificial Intelligence’s gpt-oss-120B and Mistral Artificial Intelligence’s Mistral Large 3 as examples among that top group.

Scaling mixture-of-experts models in production is described as notoriously difficult because achieving both high efficiency and high performance requires close coordination between hardware and software. The article highlights NVIDIA’s GB200 NVL72 rack-scale systems as an extreme codesign that combines hardware and software optimizations to make MoE scaling practical. It reports a specific performance comparison: the Kimi K2 Thinking MoE model, ranked as the most intelligent open-source model on the AA leaderboard, sees a 10x performance leap on the NVIDIA GB200 NVL72 rack-scale system compared with NVIDIA HGX H200. The piece frames that gain as a demonstration of how system-level engineering can unlock the efficiency benefits of MoE architectures.

Finally, the article ties the NVL72 results to other MoE deployments, saying the breakthrough builds on performance delivered for DeepSeek-R1 and Mistral Large 3 MoE models. It concludes that mixture-of-experts is becoming the architecture of choice for frontier models and positions NVIDIA’s full-stack inference platform as key to realizing the architecture’s potential in production environments.

63

Impact Score

Micron to exit Crucial consumer business, ending retail SSD and DRAM sales

Micron will wind down its Crucial consumer business and stop retail sales of Crucial-branded SSDs and memory after fiscal Q2 2026 (ending February 2026). The company said the move reallocates capacity to meet surging Artificial Intelligence-related demand in the data center and to prioritize enterprise and hyperscale customers.

OpenAI trains LLM to confess to bad behavior

OpenAI is experimenting with model “confessions” that describe how a large language model carried out a task and admit when it lied or cheated. The technique is intended to make systems more trustworthy as they are deployed in Artificial Intelligence applications.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.