at AWS re:Invent, NVIDIA and Amazon Web Services expanded a strategic collaboration with new integrations across interconnect technology, cloud infrastructure, open models and physical Artificial Intelligence. as part of that expansion, AWS will support NVIDIA NVLink Fusion, described as a platform for custom Artificial Intelligence infrastructure, to deploy AWS custom-designed silicon including next-generation Trainium4 chips for inference and agentic Artificial Intelligence model training, Graviton CPUs for a broad range of workloads and the Nitro System virtualization infrastructure.
using NVIDIA NVLink Fusion, AWS will combine NVIDIA NVLink scale-up interconnect and the NVIDIA MGX rack architecture with AWS custom silicon to increase performance and accelerate time to market for its next-generation cloud-scale Artificial Intelligence capabilities. AWS is designing Trainium4 to integrate with NVLink and NVIDIA MGX, and the companies describe this as the first of a multigenerational collaboration between NVIDIA and AWS for NVLink Fusion. the article notes that AWS has already deployed MGX racks at scale with NVIDIA GPUs.
the integration of NVLink Fusion is presented as a way for AWS to further simplify deployment and systems management across its platforms. the announcement frames the work as broadening the existing full-stack partnership so that AWS can combine NVIDIA interconnect and rack designs with its own silicon and virtualization technologies to support a range of workloads and next-generation cloud-scale Artificial Intelligence deployments.
