Huawei Ascend 950 artificial intelligence accelerator revealed with custom HBM

Huawei has revealed its Ascend 950 artificial intelligence accelerator family, pairing in-house HBM memory with a focus on large-scale system performance rather than single-chip dominance.

Huawei’s next-generation Ascend 950 artificial intelligence accelerator has been shown in detail for the first time, highlighting the company’s custom silicon and high bandwidth memory. The design combines Huawei’s first self-developed HBM with a new generation of artificial intelligence acceleration, signaling a strategy built around vertically integrated components and system-level optimization.

The Ascend 950 family is planned for early 2026 and will launch with two variants. The 950PR model includes 128 GB of in-house HBM with around 1.6 TB/s of bandwidth, while the 950DT model increases memory to 144 GB and boosts bandwidth to nearly 4 TB/s. Both chips target one PetaFLOP of FP8, and two PetaFLOPS of FP4, positioning them as high-performance accelerators even though Huawei currently trails NVIDIA in per-chip performance.

Huawei’s competitive approach centers on dense packaging and aggressive networking across clusters, aiming to win through scale rather than relying solely on raw per-chip metrics. Although the company has not confirmed the manufacturing process, the accelerator will likely use SMIC’s newest N+3 node with 5 nm-class features. Chinese SMIC has officially achieved volume production of its newest 5 nm node relying on the deep ultraviolet to manufacture its silicon, and since the first customer for N+3 was Huawei with the Kirin 9030 SoC, it is seen as logical that the Ascend artificial intelligence accelerator family would also be produced on the same node.

64

Impact Score

Indiana launches Artificial Intelligence business portal

Indiana is rolling out IN AI, a statewide portal meant to help employers adopt Artificial Intelligence with practical guidance, workshops and peer support. State leaders and business groups are positioning the effort as a way to raise productivity, wages and job growth while keeping workers at the center.

Goodfire launches model debugging tool for large language models

Goodfire has introduced Silico, a mechanistic interpretability platform designed to let developers inspect and adjust model behavior during development. The company is positioning it as a way to give smaller teams deeper control over open-source models and more trustworthy outputs.

Nvidia launches nemotron 3 nano omni for enterprise agents

Nvidia has introduced Nemotron 3 Nano Omni, a multimodal open model designed to support enterprise agents that reason across vision, speech and language. The launch extends Nvidia’s push beyond hardware into models and services while targeting more efficient agentic workflows.

Intel 18A-P node improves performance and efficiency

Intel plans to present new results for its 18A-P process at the VLSI 2026 Symposium, highlighting gains in performance, power efficiency, and manufacturing predictability. The updated node is positioned as a stronger option for customers seeking 18A density with better operating characteristics.

EA CEO defends broader Artificial Intelligence use in game development

EA CEO Andrew Wilson defended the company’s internal use of Artificial Intelligence after employee claims that the tools were slowing work rather than helping. He framed the technology as an aid for repetitive quality assurance tasks, even as concerns persist over its broader impact on development.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.