AMD challenges Nvidia on software and CPUs

AMD is pressing Nvidia on two fronts: reducing lock-in around gpu software and defending its lead in server cpus as Nvidia expands with Grace and Vera. The contest is shaping around open developer tools, inference performance, and control of Artificial Intelligence data center orchestration.

AMD is escalating its competition with Nvidia across both gpu software and server cpus, aiming to weaken CUDA lock-in while protecting its position in processor infrastructure. Ahead of Nvidia’s GTC event, AMD pushed back on claims that it trails badly in inference software, with Anush Elangovan, AMD’s VP for Artificial Intelligence software, arguing that updates to the ROCm stack have narrowed the gap and in some cases surpassed Nvidia’s B200 on four-bit floating point (FP4). He also argued that customer demand remains centered on eight-bit floating point (FP8), saying “most” users operate there and that “FP8 is still king of the hill.”

AMD is positioning ROCm as an open alternative to CUDA for Artificial Intelligence and high-performance computing workloads, emphasizing compatibility with frameworks such as PyTorch, TensorFlow, and JAX. A core part of that strategy is heavier investment in Triton, the open, Python-first gpu compiler originally developed at OpenAI. AMD sees Triton as the highest-level abstraction layer for gpu programming and wants it to become “the de facto” standard so moving workloads from Nvidia to AMD is “zero friction.” Its own lower-level tools, including Fly DSL and Wave, are intended to handle hardware-specific tuning underneath that abstraction. AMD also frames Nvidia’s CUDA Tile push as a reaction to Triton’s growing role in democratizing gpu programming.

The rivalry is also expanding into the cpu layer, where AMD is trying to defend what it sees as an established advantage as Nvidia promotes Grace and Vera. Days before GTC, AMD highlighted benchmark claims for its server chips, citing SPEC CPU Benchmark data that its 5th-Gen EPYC CPU offered 2.1-times higher performance per core against Nvidia’s Grace Superchip systems, while up to a 2.26-times uplift in operations per watt. AMD is also tying cpus more directly to emerging agentic Artificial Intelligence workloads, arguing that processors act as the control plane for gpu-heavy data centers by orchestrating work and managing more complex tasks.

Cloud providers are already reflecting that positioning. Microsoft added AMD’s Turin processors to its Da/Ea/Fasv7-series virtual machines and said the chips deliver 35% better CPU performance compared to the prior v6 AMD-based generation, along with higher instructions-per-clock, greater memory bandwidth, and support for advanced vector instructions. Google Cloud has also adopted AMD’s 5th-Gen EPYC processors for its C4D and H4D instances aimed at Artificial Intelligence inference, high-performance computing, and general-purpose workloads. AMD has further extended that push with edge-focused cpu variants launched last September for latency-critical applications, signaling that the company intends to compete with Nvidia from software stack to server control plane.

55

Impact Score

Nvidia’s hold on the Artificial Intelligence boom

Nvidia is portrayed as a central power broker in the Artificial Intelligence industry, with Jensen Huang’s remarks underscoring the company’s influence. The available details point to a chip giant seen as a kingmaker in the market.

NC State researchers target safer large language models

North Carolina State University researchers developed a framework for understanding why large language models can produce unsafe outputs and identified neuron-level components tied to safety decisions. Their approach aims to preserve safety during fine-tuning while reducing the performance costs of alignment.

What comes next for large language models and agents

Google and Nvidia researchers outlined a near-term future in which large language models and agents act more autonomously, learn continuously, and operate at machine speed. They also pointed to new roles in chip design, robotics, cybersecurity, and education.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.