Alibaba Cloud expands elastic GPU instance portfolio for artificial intelligence and graphics workloads

Alibaba Cloud has detailed a broad lineup of elastic GPU instance families across gn, vgn, and sgn series, targeting artificial intelligence training and inference, high performance computing, and professional graphics workloads with options from virtual GPUs to confidential computing.

Alibaba Cloud has outlined an extensive range of Elastic GPU Service instance families across the gn, vgn, and sgn series, designed to combine GPU and CPU resources for artificial intelligence, high performance computing, and graphics workloads. The platform supports both GPU-accelerated and vGPU-accelerated instance types, largely built on the third-generation SHENLONG architecture, with I/O optimized designs, Non-Volatile Memory Express (NVMe) support, and dual-stack IPv4 and IPv6 networking. Many of the families integrate NVIDIA GRID Virtual Workstation licenses for certified CAD and professional graphics acceleration, and can be used as relatively lightweight GPU-accelerated compute instances for small-scale artificial intelligence inference.

The sgn8ia and sgn7i-vws families use virtual GPUs and shared CPUs to drive concurrent artificial intelligence reasoning and 3D graphics workloads such as remote design and cloud gaming, with sgn8ia relying on NVIDIA Lovelace GPUs and AMD Genoa processors that deliver a clock speed of 3.4 GHz to 3.75 GHz. The sgn7i-vws and vgn7i-vws families pair NVIDIA A10 GPUs with 2.9 GHz Intel Xeon Scalable Ice Lake processors and offer fine-grained GPU slicing, where configurations such as “NVIDIA A10 × 1/12” indicate that a single GPU is partitioned into 12 vGPU segments. The vgn6i-vws family upgrades earlier vgn6i instances to newer GRID drivers on NVIDIA T4 GPUs, supporting 1/4 and 1/2 GPU capacity slices, 4 GB and 8 GB GPU memory options, and a CPU-to-memory ratio of 1:5, with use cases spanning cloud gaming rendering, augmented and virtual reality, and artificial intelligence inference in elastic internet environments.

At the high end, gn8v and gn8v-tee represent Alibaba Cloud’s 8th-generation GPU-accelerated compute-optimized instances, aimed at ultra large language model training and inference, autonomous driving training, and multi-GPU parallel inference on models with more than 70 billion parameters. In the described scenarios, traditional artificial intelligence model training and autonomous driving training use GPUs that deliver computing power of up to 39.5 TFLOPS in the single-precision FP32 format, with each GPU equipped with 96 GB of HBM3 memory and up to 4 TB/s of memory bandwidth, and interconnected by 900 GB/s NVLink links for multi-GPU efficiency gains. The gn8v-tee variants add end-to-end confidential computing using Intel Trust Domain Extensions and NVIDIA confidential computing to secure model and inference data, while all gn8v-class instances adopt CIPU 1.0, 4th-generation Intel Xeon Scalable processors with base frequencies of up to 2.8 GHz and all-core turbo frequencies of up to 3.1 GHz, and network performance that can reach a packet forwarding rate of up to 30,000,000 pps on 8-GPU configurations.

The gn8is family targets artificial intelligence generated content workloads with NVIDIA L20 GPUs, each providing 48 GB of memory and FP8 support, and is positioned for inference on language models with fewer than 70 billion parameters. Earlier gn7e, gn7i, gn7s, gn7, gn6i, gn6e, gn6v, gn5, and gn5i families cover a wide spectrum of training and inference tasks, including deep learning for image classification and autonomous vehicles, high performance scientific computing, cloud gaming, and multimedia encoding. These lines span NVIDIA A10, A30, T4, V100, P100, and P4 accelerators, offer differing CPU-to-memory ratios, and progressively increase network bandwidth and packet forwarding rates through larger instance sizes. Across the portfolio, Alibaba Cloud emphasizes predictable performance from the SHENLONG architecture, flexible combinations of GPU counts and vCPU resources, and broad block storage options such as ESSDs, ESSD AutoPL disks, Regional ESSDs, and in some older families, local NVMe SSDs, to match a range of artificial intelligence, graphics, and compute-intensive workloads.

55

Impact Score

Apple creator studio bundles pro creative apps and artificial intelligence features

Apple creator studio is a new subscription bundle that combines Final Cut Pro, Pixelmator Pro, Logic Pro and enhanced productivity features in Keynote, Pages, Numbers and Freeform, with artificial intelligence tools layered on top of Apple Intelligence. The service launches on January 28 and offers different pricing for standard users and education customers, plus trial periods for new subscribers and recent device buyers.

Generative artificial intelligence for uk corporate tax: use cases, risks, and controls

Generative artificial intelligence is reshaping corporate tax workflows in the uk by accelerating drafting, research, and knowledge reuse, while elevating the importance of professional judgement, governance, and data protection. Firms are moving from experimentation toward structured adoption as compliance pressures, software based filing, and expectations for auditability increase.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.