OpenAI builds transparent weight-sparse transformer to reveal how Artificial Intelligence models work

OpenAI has built an experimental, more transparent large language model that helps researchers trace how Artificial Intelligence systems compute. The model is smaller and slower than commercial products but offers clearer internal mechanisms for study.

OpenAI has developed an experimental large language model called a weight-sparse transformer that is far easier to analyze than typical models. The project responds to a core problem: modern models are black boxes, and researchers cannot fully explain why they hallucinate or fail when applied to important domains. Leo Gao, a research scientist at OpenAI, said the work aims to improve safety as Artificial Intelligence systems are integrated into high-stakes tasks.

The research sits in the field of mechanistic interpretability, which tries to map the internal circuits models use to carry out tasks. Most existing models use dense neural networks, where neurons connect broadly and learned features are spread across many units. That structure creates superposition, where individual neurons represent multiple features, making it hard to attribute behavior to specific parts. OpenAI instead constructed a weight-sparse transformer in which each neuron connects to only a few others, forcing features into localized clusters and making it easier to relate neurons or groups of neurons to concrete concepts and functions.

The resulting model is much smaller and slower than leading commercial models and at most as capable as GPT-1, according to Gao. OpenAI has used it to trace exact chains of computation for simple tasks, such as adding a matching quotation mark to a block of text, and identified a learned circuit that mirrors an algorithm one might implement by hand. External researchers praised the approach as promising, while others warned it may not scale to larger, more capable models. Gao and Dan Mossing of OpenAI acknowledge the limitations but say the technique could eventually yield a fully interpretable model on the order of GPT-3, which would provide deep insight into how complex Artificial Intelligence systems function.

58

Impact Score

Adobe plans outcome-based pricing for Artificial Intelligence agents

Adobe is positioning its Artificial Intelligence agents around performance-based pricing, charging only when the software completes useful work. The approach points to a more results-oriented model for selling generative Artificial Intelligence tools to business customers.

Tech firms commit billions to Artificial Intelligence infrastructure

Amazon, OpenAI, Nvidia, Meta, Google and others are signing increasingly large cloud, chip and data center agreements as demand for Artificial Intelligence infrastructure accelerates. The latest wave of deals spans investments, compute purchases, chip supply agreements and data center buildouts.

JEDEC outlines LPDDR6 expansion for data centers

JEDEC has previewed planned updates to LPDDR6 aimed at pushing the memory standard beyond mobile devices and into selected data center and accelerated computing use cases. The roadmap includes higher-capacity packaging options, flexible metadata support, 512 GB densities, and a new SOCAMM2 module standard.

Tsmc debuts A13 process technology

Tsmc has introduced its A13 process at its 2026 North America Technology Symposium as a tighter version of A14 aimed at next-generation Artificial Intelligence, high performance computing, and mobile designs. The company positions the node as a more compact and efficient option with backward-compatible design rules for faster migration.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.