Google readies TPUv8ax for training and TPUv8x for inference workloads

Google is preparing its eighth generation of tensor processing units with separate TPUv8ax and TPUv8x chips tuned for large Artificial Intelligence training and inference tasks, while shifting more design work in-house and relying selectively on Broadcom and MediaTek.

Google is expanding its custom Artificial Intelligence infrastructure with two new tensor processing unit variants targeting distinct workloads. The eighth generation lineup introduces TPUv8ax, codenamed ‘Sunfish’, which is oriented toward training large Artificial Intelligence models such as Gemini, and TPUv8x, codenamed ‘Zebrafish’, which is tailored for large scale inference deployments. The new designs are intended to strengthen Google’s internal cloud hardware stack as external interest in its custom accelerators grows.

For the training focused TPUv8ax ‘Sunfish’, Google has partnered with Broadcom and its custom design group, which is responsible for end to end design, memory integration, supporting hardware, and packaging. Broadcom delivers a finished TPU product that can be dropped into Google’s existing server infrastructure, reducing integration friction for new training capacity. This arrangement keeps Broadcom deeply involved in the most complex parts of the chip development process while allowing Google to standardize deployment across its data centers.

The inference oriented TPUv8x ‘Zebrafish’ uses a different collaboration model that shifts more responsibility to Google. For ‘Zebrafish’, Google has brought in MediaTek in a limited role, while Google sources wafers and memory directly from suppliers and uses MediaTek mainly for supporting chips and packaging, where Google has less experience. This approach means a larger share of chip design work now occurs in house, which reduces dependence on external partners but still acknowledges gaps in Google’s full stack chip design capabilities. Specific performance and memory figures for TPUv8ax and TPUv8x have not been disclosed, but expectations are that they will exceed the TPUv7 ‘Ironwood’, which carries 4,614 TeraFLOPS at FP8 precision and 192 GB of HBM memory.

58

Impact Score

Indiana launches Artificial Intelligence business portal

Indiana is rolling out IN AI, a statewide portal meant to help employers adopt Artificial Intelligence with practical guidance, workshops and peer support. State leaders and business groups are positioning the effort as a way to raise productivity, wages and job growth while keeping workers at the center.

Goodfire launches model debugging tool for large language models

Goodfire has introduced Silico, a mechanistic interpretability platform designed to let developers inspect and adjust model behavior during development. The company is positioning it as a way to give smaller teams deeper control over open-source models and more trustworthy outputs.

Nvidia launches nemotron 3 nano omni for enterprise agents

Nvidia has introduced Nemotron 3 Nano Omni, a multimodal open model designed to support enterprise agents that reason across vision, speech and language. The launch extends Nvidia’s push beyond hardware into models and services while targeting more efficient agentic workflows.

Intel 18A-P node improves performance and efficiency

Intel plans to present new results for its 18A-P process at the VLSI 2026 Symposium, highlighting gains in performance, power efficiency, and manufacturing predictability. The updated node is positioned as a stronger option for customers seeking 18A density with better operating characteristics.

EA CEO defends broader Artificial Intelligence use in game development

EA CEO Andrew Wilson defended the company’s internal use of Artificial Intelligence after employee claims that the tools were slowing work rather than helping. He framed the technology as an aid for repetitive quality assurance tasks, even as concerns persist over its broader impact on development.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.