NVIDIA and AWS expand full-stack partnership for Artificial Intelligence infrastructure

At AWS re:Invent, NVIDIA and Amazon Web Services expanded a strategic collaboration to integrate interconnect technology, cloud infrastructure, open models and physical Artificial Intelligence, with AWS adding support for NVIDIA NVLink Fusion.

at AWS re:Invent, NVIDIA and Amazon Web Services expanded a strategic collaboration with new integrations across interconnect technology, cloud infrastructure, open models and physical Artificial Intelligence. as part of that expansion, AWS will support NVIDIA NVLink Fusion, described as a platform for custom Artificial Intelligence infrastructure, to deploy AWS custom-designed silicon including next-generation Trainium4 chips for inference and agentic Artificial Intelligence model training, Graviton CPUs for a broad range of workloads and the Nitro System virtualization infrastructure.

using NVIDIA NVLink Fusion, AWS will combine NVIDIA NVLink scale-up interconnect and the NVIDIA MGX rack architecture with AWS custom silicon to increase performance and accelerate time to market for its next-generation cloud-scale Artificial Intelligence capabilities. AWS is designing Trainium4 to integrate with NVLink and NVIDIA MGX, and the companies describe this as the first of a multigenerational collaboration between NVIDIA and AWS for NVLink Fusion. the article notes that AWS has already deployed MGX racks at scale with NVIDIA GPUs.

the integration of NVLink Fusion is presented as a way for AWS to further simplify deployment and systems management across its platforms. the announcement frames the work as broadening the existing full-stack partnership so that AWS can combine NVIDIA interconnect and rack designs with its own silicon and virtualization technologies to support a range of workloads and next-generation cloud-scale Artificial Intelligence deployments.

68

Impact Score

Samsung completes hbm4 development, awaits NVIDIA approval

Samsung says it has cleared Production Readiness Approval for its first sixth-generation hbm (hbm4) and has shipped samples to NVIDIA for evaluation. Initial samples have exceeded NVIDIA’s next-gen GPU requirement of 11 Gbps per pin and hbm4 promises roughly 60% higher bandwidth than hbm3e.

NVIDIA and AWS expand full-stack partnership for Artificial Intelligence compute platform

NVIDIA and AWS expanded integration around Artificial Intelligence infrastructure at AWS re:Invent, announcing support for NVIDIA NVLink Fusion with Trainium4, Graviton and the Nitro System. the move aims to unify NVIDIA scale-up interconnect and MGX rack architecture with AWS custom silicon to speed cloud-scale Artificial Intelligence deployments.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.