AMD Instinct MI350 platform for Artificial Intelligence and high-performance computing on GIGABYTE servers

The AMD Instinct MI350 Series, launched in June 2025, brings 4th Gen AMD CDNA architecture and TSMC 3nm process to data center workloads, with 288 GB HBM3E and up to 8 TB/s memory bandwidth. GIGABYTE pairs these accelerators and the MI300 family with 8-GPU UBB servers, direct liquid cooling options, and ROCm 7.0 software support for large-scale Artificial Intelligence and high-performance computing deployments.

The AMD Instinct MI350 Series, announced in June 2025, is presented as a major generational step for data center computing. Built on 4th Gen AMD CDNA architecture and fabricated with TSMC’s 3nm process, the MI350 family includes MI350X and MI355X GPUs with 288 GB of HBM3E memory and up to 8 TB/s bandwidth. AMD and GIGABYTE materials highlight up to 4x generational Artificial Intelligence compute improvement and a claimed 35x uplift in inference performance for certain workloads, positioning the MI350 series for training large models, high-speed inference, and demanding HPC tasks.

GIGABYTE has integrated the AMD Instinct family into a portfolio of servers optimized for density, cooling, and scale. The vendor offers UBB-based 8-GPU configurations and OAM modules such as MI325X and MI300X on universal baseboards inside G-series systems. Product examples include 4U liquid-cooled designs and 8U air-cooled enclosures including the G4L3 and G893 series, plus multi-socket systems such as the G383 series that host the MI300A APU in a four-LGA-socket configuration. Cooling options range from passive to direct liquid cooling, and systems support PCIe Gen5, Infinity Fabric links, and virtualization partitioning to address varied Artificial Intelligence and high-performance computing requirements.

The software stack emphasizes AMD ROCm 7.0 and ecosystem readiness. ROCm 7 is described as providing day-0 framework support for PyTorch, TensorFlow, JAX, ONNX Runtime, Triton, and vLLM, plus transformer kernels and distributed inference optimizations for OCP-FP8/MXFP formats. AMD testing cited from May 15, 2025 reports ROCm 7 preview delivering up to 3.5x faster inference and 3x faster training versus ROCm 6 on an eight-MI300X configuration across Llama 3.1-70B, Qwen 72B, and Deepseek-R1 models. Together, GIGABYTE hardware and AMD accelerators are framed as an open, high-density solution for enterprises and research centers pursuing large-scale Artificial Intelligence and HPC deployments.

68

Impact Score

Yann LeCun world model startup challenges OpenAI dominance

Yann LeCun’s new world model venture, Advanced Machine Intelligence Labs, is raising massive early funding to pursue a physics-grounded alternative to large language models, directly challenging OpenAI’s text-centric strategy and market position.

Samsung SOCAMM2 LPDDR module targets next generation artificial intelligence data centers

Samsung has introduced SOCAMM2, an LPDDR based server memory module with a modular, detachable design, aimed at improving bandwidth, power efficiency, and integration in artificial intelligence data centers. The company is already supplying customer samples as demand rises for low power memory tailored to continuous artificial intelligence workloads.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.