Nvidia boosts Vera Rubin bandwidth ahead of AMD Instinct MI400 launch

Nvidia has repeatedly raised memory bandwidth targets for its Vera Rubin VR200 NVL72 system to stay competitive with AMD’s Instinct MI400 accelerators, as both companies escalate their Artificial Intelligence and high performance computing roadmaps.

Nvidia has been iterating on its upcoming Vera Rubin superchip as it prepares for first server shipments in late summer, with the company significantly increasing the memory bandwidth of the VR200 NVL72 system over the past year to better match AMD’s Instinct MI400 accelerator family. According to SemiAnalysis, Nvidia’s initial target for the ‘Vera Rubin’ VR200 NVL72 system was 13 TB/s in March, which was upgraded to 20.5 TB/s by September. At CES 2026, Nvidia confirmed that the VR200 NVL72 system is now operating at 22 TB/s of bandwidth, bringing it ahead of AMD’s Instinct MI455X accelerator, which has 19.6 TB/s, after originally trailing in system bandwidth. Nvidia is reported to have closed this gap by adopting faster DRAM and refining interconnects that link CPUs, GPUs, and the broader system.

AMD has been positioning its Instinct MI400 lineup directly against Nvidia’s Vera Rubin, emphasizing advantages in memory capacity and scale-out connectivity while targeting comparable compute throughput and bandwidth. In November, AMD compared the MI400 lineup to Nvidia’s upcoming ‘Vera Rubin’ series, claiming similar compute performance and memory bandwidth but 1.5 times higher memory capacity and scale-out bandwidth. The company is framing MI400 as a generational leap over its current MI350 accelerators, with a particular focus on large scale Artificial Intelligence workloads and dense high performance computing.

On the raw performance side, AMD claims it will deliver up to 40 FP4 and 20 FP8 PFLOPs, roughly twice the compute performance of the current MI350. The GPUs also transition to HBM4 memory from HBM3e, increasing capacity from 288 GB to 432 GB and raising total bandwidth from 8 TB/s to 19.6 TB/s. Each GPU provides 300 GB/s of scale-out bandwidth and adds broader Artificial Intelligence data format support along with expanded Artificial Intelligence pipelines, underscoring AMD’s focus on training and inference efficiency. AMD is planning two key models in the series: the Instinct MI455X, aimed at large-scale Artificial Intelligence training and inference deployments, and the MI430X, targeted at high performance computing and Artificial Intelligence, setting up a direct competitive clash with Nvidia’s Vera Rubin platforms as both enter the market.

55

Impact Score

Indiana launches Artificial Intelligence business portal

Indiana is rolling out IN AI, a statewide portal meant to help employers adopt Artificial Intelligence with practical guidance, workshops and peer support. State leaders and business groups are positioning the effort as a way to raise productivity, wages and job growth while keeping workers at the center.

Goodfire launches model debugging tool for large language models

Goodfire has introduced Silico, a mechanistic interpretability platform designed to let developers inspect and adjust model behavior during development. The company is positioning it as a way to give smaller teams deeper control over open-source models and more trustworthy outputs.

Nvidia launches nemotron 3 nano omni for enterprise agents

Nvidia has introduced Nemotron 3 Nano Omni, a multimodal open model designed to support enterprise agents that reason across vision, speech and language. The launch extends Nvidia’s push beyond hardware into models and services while targeting more efficient agentic workflows.

Intel 18A-P node improves performance and efficiency

Intel plans to present new results for its 18A-P process at the VLSI 2026 Symposium, highlighting gains in performance, power efficiency, and manufacturing predictability. The updated node is positioned as a stronger option for customers seeking 18A density with better operating characteristics.

EA CEO defends broader Artificial Intelligence use in game development

EA CEO Andrew Wilson defended the company’s internal use of Artificial Intelligence after employee claims that the tools were slowing work rather than helping. He framed the technology as an aid for repetitive quality assurance tasks, even as concerns persist over its broader impact on development.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.