Mira Murati’s Thinking Machines solves LLM nondeterminism

Mira Murati’s startup Thinking Machines says it has eliminated nondeterminism in large language model outputs by introducing batch-invariant kernels. The approach yields identical responses at temperature zero, promising stronger reproducibility for research, audits, and safety-critical Artificial Intelligence.

Mira Murati’s new company, Thinking Machines, claims a fix for one of the most persistent issues in large language model behavior: nondeterministic outputs. The team says it has engineered batch-invariant kernels for key operations so that models return identical responses when run at temperature zero. By removing variation tied to how inputs are grouped or processed, the method is positioned as a way to deliver deterministic outputs on demand, a foundational requirement for trustworthy Artificial Intelligence systems.

The company frames determinism as a practical necessity rather than a theoretical nicety. Nondeterministic behavior makes it difficult to reproduce results, verify model changes, or maintain consistent behavior across runs. Thinking Machines’ batch-invariant kernels are presented as a direct solution to this, asserting that the same prompt will yield the same answer at temperature zero regardless of batch composition. That consistency can simplify debugging, streamline comparisons during model iteration, and make it easier to establish a stable baseline for evaluation.

According to the company’s positioning, the benefits are especially pronounced in research, audits, and safety-critical Artificial Intelligence. In research workflows, deterministic outputs enable repeatable experiments and clearer attribution of performance changes to specific interventions. For audits and compliance, guaranteed identical responses support traceability and verifiable record keeping. In safety-critical contexts, where predictability is paramount, identical outputs at temperature zero help reduce operational risk and strengthen the case for deployment under strict reliability constraints.

Thinking Machines characterizes this advance as a potential shift in how the industry approaches inference standards. If batch-invariant kernels become a reference approach, deterministic, temperature-zero inference could move from a best effort practice to an expected baseline for production systems. The company suggests that the breakthrough may redefine expectations for deterministic, trustworthy Artificial Intelligence, signaling a move toward systems that are more reproducible, easier to audit, and better aligned with safety and governance requirements.

70

Impact Score

Artificial Intelligence LLM confessions and geothermal hot spots

OpenAI is testing a method that prompts large language models to produce confessions explaining how they completed tasks and acknowledging misconduct, part of efforts to make multitrillion-dollar Artificial Intelligence systems more trustworthy. Separately, startups are using Artificial Intelligence to locate blind geothermal systems and energy observers note seasonal patterns in nuclear reactor operations.

Saudi Artificial Intelligence startup launches Arabic LLM

Misraj Artificial Intelligence unveiled Kawn, an Arabic large language model, at AWS re:Invent and launched Workforces, a platform for creating and managing Artificial Intelligence agents for enterprises and public institutions.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.