Mira Murati’s Thinking Machines solves LLM nondeterminism

Mira Murati’s startup Thinking Machines says it has eliminated nondeterminism in large language model outputs by introducing batch-invariant kernels. The approach yields identical responses at temperature zero, promising stronger reproducibility for research, audits, and safety-critical Artificial Intelligence.

Mira Murati’s new company, Thinking Machines, claims a fix for one of the most persistent issues in large language model behavior: nondeterministic outputs. The team says it has engineered batch-invariant kernels for key operations so that models return identical responses when run at temperature zero. By removing variation tied to how inputs are grouped or processed, the method is positioned as a way to deliver deterministic outputs on demand, a foundational requirement for trustworthy Artificial Intelligence systems.

The company frames determinism as a practical necessity rather than a theoretical nicety. Nondeterministic behavior makes it difficult to reproduce results, verify model changes, or maintain consistent behavior across runs. Thinking Machines’ batch-invariant kernels are presented as a direct solution to this, asserting that the same prompt will yield the same answer at temperature zero regardless of batch composition. That consistency can simplify debugging, streamline comparisons during model iteration, and make it easier to establish a stable baseline for evaluation.

According to the company’s positioning, the benefits are especially pronounced in research, audits, and safety-critical Artificial Intelligence. In research workflows, deterministic outputs enable repeatable experiments and clearer attribution of performance changes to specific interventions. For audits and compliance, guaranteed identical responses support traceability and verifiable record keeping. In safety-critical contexts, where predictability is paramount, identical outputs at temperature zero help reduce operational risk and strengthen the case for deployment under strict reliability constraints.

Thinking Machines characterizes this advance as a potential shift in how the industry approaches inference standards. If batch-invariant kernels become a reference approach, deterministic, temperature-zero inference could move from a best effort practice to an expected baseline for production systems. The company suggests that the breakthrough may redefine expectations for deterministic, trustworthy Artificial Intelligence, signaling a move toward systems that are more reproducible, easier to audit, and better aligned with safety and governance requirements.

70

Impact Score

Generative Artificial Intelligence in travel

PhocusWire’s hub compiles in-depth reporting on generative Artificial Intelligence across the travel industry, from customer service and marketing to product development. The page highlights new tools, research and leaders shaping automation, personalization and decision-making.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.