NVIDIA Blackwell Dominates MLPerf Inference Benchmarks

NVIDIA Blackwell´s performance in MLPerf Inference V5.0 sets new records, showcasing cutting-edge capabilities in Artificial Intelligence.

In the latest MLPerf Inference V5.0 benchmarks, NVIDIA´s Blackwell platform has set new records, marking a significant achievement in artificial intelligence inference capabilities. For the first time, NVIDIA used its GB200 NVL72 system, a rack-scale solution designed for AI reasoning, for the submission. This system effectively connects 72 NVIDIA Blackwell GPUs into a single massive GPU, achieving up to 30x higher throughput on the Llama 3.1 405B benchmark compared to previous submissions.

Designed for AI factories, the NVIDIA Blackwell platform demonstrates the future of data processing by transforming raw data into real-time insights. The emphasis is on delivering accurate and swift responses to queries at minimal costs to numerous users. Innovations across technology stacks in silicon, network systems, and software are pushing the boundaries of AI capabilities, enabling smarter models with billions of parameters to efficiently deliver insights in real-time, while managing costs and computing resources.

Apart from Blackwell, the NVIDIA Hopper platform also displayed exceptional performance, significantly improving throughput on the Llama 2 70B benchmark due to full-stack optimizations. This ongoing enhancement in NVIDIA´s platforms underscores the sustained value and adaptability of its AI solutions amidst growing model complexities and demand for responsive user experiences. The inclusive participation of several partners and rigorous peer-reviewed benchmarking further highlights the comprehensive reach and influence of NVIDIA´s evolving AI technologies.

78

Impact Score

Artificial Intelligence speeds quantum encryption threat timeline

Research from Google and Oratomic suggests quantum computers capable of breaking core internet encryption may arrive sooner than expected. Artificial Intelligence played a key role in improving one of the new algorithms, raising fresh urgency around post-quantum security.

New methods aim to improve Large Language Model reasoning

A new study on arXiv outlines algorithmic techniques designed to strengthen Large Language Model reasoning and reduce hallucinations. The work reports better logical consistency and stronger performance on mathematical and coding benchmarks.

Nvidia acquisition of SchedMD raises Slurm neutrality concerns

Nvidia’s purchase of SchedMD has given it control of Slurm, an open-source scheduler that sits at the center of many supercomputing and large-model training systems. Researchers and engineers are watching for signs that support could tilt toward Nvidia hardware over AMD and Intel alternatives.

Mustafa Suleyman says Artificial Intelligence compute growth is still accelerating

Mustafa Suleyman argues that Artificial Intelligence development is being propelled by simultaneous advances in chips, memory, networking, and software efficiency rather than nearing a hard limit. He contends that rising compute capacity and falling deployment costs will push systems beyond chatbots toward more capable agents.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.