NVIDIA brings CUDA support to RISC-V, embracing open-source ISA

NVIDIA brings its CUDA platform to RISC-V, signaling confidence in the open-source architecture and paving the way for future heterogeneous systems in artificial intelligence and high-performance computing.

At the RISC-V Summit in China, NVIDIA officially revealed that its flagship CUDA software library has been ported to the open-source RISC-V instruction set architecture. Frans Sijstermans, NVIDIA´s vice president of hardware engineering, announced the milestone, underscoring the company´s belief that RISC-V is quickly gaining traction across the hardware landscape. Historically, NVIDIA has strategically adapted CUDA to a variety of instruction sets, including x86, Arm, PowerPC, and SPARC, positioning itself to support GPU acceleration on any competitive enterprise platform from the outset. Extending full CUDA capabilities to RISC-V indicates NVIDIA´s readiness to back emerging architectures before widespread enterprise adoption.

NVIDIA´s architectural vision places its GPUs at the epicenter of workload acceleration, with supporting RISC-V CPUs responsible for managing CUDA drivers, application workflows, and operating system functions. This approach enables seamless orchestration of parallel computing tasks entirely within the CUDA environment. A demonstration at the summit showcased this synergy: RISC-V processors together with NVIDIA GPUs and a dedicated data processing unit (DPU) combine to form a unified platform encompassing compute, control, and data movement capabilities. Notably, NVIDIA already deploys NV-RISC-V microcontrollers for internal GPU control logic, highlighting a longstanding commitment to integrating RISC-V into its hardware ecosystem.

With CUDA now natively supported on RISC-V, NVIDIA is in a position to consider future processors—potentially including successors to its Grace CPU—built entirely on this open-source instruction set. As RISC-V encroaches on the server market, propelled by the RVA23 specification (a requirement for NVIDIA´s CUDA compatibility), new opportunities emerge for heterogeneous system designs that mix open-source and proprietary solutions. The move is poised to accelerate both enterprise and research adoption of RISC-V, further blurring traditional boundaries between instruction sets in high-performance computing and artificial intelligence deployments.

79

Impact Score

Tech firms commit billions to Artificial Intelligence infrastructure

Amazon, OpenAI, Nvidia, Meta, Google and others are signing increasingly large cloud, chip and data center agreements as demand for Artificial Intelligence infrastructure accelerates. The latest wave of deals spans investments, compute purchases, chip supply agreements and data center buildouts.

JEDEC outlines LPDDR6 expansion for data centers

JEDEC has previewed planned updates to LPDDR6 aimed at pushing the memory standard beyond mobile devices and into selected data center and accelerated computing use cases. The roadmap includes higher-capacity packaging options, flexible metadata support, 512 GB densities, and a new SOCAMM2 module standard.

Tsmc debuts A13 process technology

Tsmc has introduced its A13 process at its 2026 North America Technology Symposium as a tighter version of A14 aimed at next-generation Artificial Intelligence, high performance computing, and mobile designs. The company positions the node as a more compact and efficient option with backward-compatible design rules for faster migration.

Google unveils eighth-generation tensor processor units

Google introduced its eighth generation of custom tensor processor units with separate designs for training and inference. The new TPU 8t and TPU 8i are aimed at large-scale model training, serving, and agentic workloads.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.