Sarvam AI Unveils Flagship LLM Surpassing Language and Reasoning Benchmarks

Bengaluru’s Sarvam AI debuts its 24-billion-parameter multilingual model, outperforming global leaders in Indian languages, math, and code for advanced Artificial Intelligence applications.

Bengaluru-based Sarvam AI, a start-up selected under the Indian government´s IndiaAI Mission, has introduced its flagship Large Language Model (LLM) named Sarvam-M. The new model is a 24-billion-parameter, multilingual, hybrid-reasoning, text-only solution built upon Mistral Small—an open-weight model from French company Mistral AI. Sarvam-M is designed for a range of applications including conversational agents, translation tools, and educational services, reflecting its broad ambitions to support India’s unique linguistic and technological landscape.

Sarvam-M sets new performance benchmarks in Indian languages, mathematical reasoning, and programming competence. The company reports that the model delivers a 20% average improvement over its base on Indian language tasks, a 21.6% enhancement in math problems, and a 17.6% uptick on coding benchmarks. Particularly striking is its +86% improvement on romanized Indian language GSM-8K math benchmarks, underscoring its strength in combining linguistic and logical reasoning relevant to local contexts. This release follows the firm’s earlier rollout of Bulbul, a speech model supporting 11 Indian languages and authentic accents.

In comparative evaluations, Sarvam-M is said to outperform Meta’s LLaMA-4 Scout and is competitive with larger models such as LLaMA-3.3 70B and Google’s Gemma 3 27B, even though those models are trained on significantly more data. While Sarvam-M exhibits slightly diminished scores (about 1% below baseline) on English knowledge benchmarks, it is made available as open source on Hugging Face, with API access for developers, promoting adaptability and research. Its architecture uniquely supports two operating modes: ´think´ mode for tasks requiring complex logic and computation, and ´non-think´ mode for efficient, general-purpose text generation and conversation. Importantly, Sarvam-M is post-trained on Indian languages alongside English, ensuring robust Indic cultural representation, and supports both native Indic scripts and romanized forms for diverse user needs.

72

Impact Score

Compression and voice models reshape Artificial Intelligence efficiency

Recent releases focused on infrastructure rather than headline model breakthroughs, with gains in compression and voice systems pointing to lower inference costs and broader deployment. Google and Mistral highlighted two distinct paths for real-time audio, while TurboQuant targeted one of the most expensive bottlenecks in long-context inference.

Judge blocks Pentagon move against Anthropic

A federal judge temporarily blocked the Pentagon from labeling Anthropic a supply chain risk after finding major gaps between public threats, legal authority, and the government’s courtroom arguments. The dispute has become a test of how far the government can go in punishing an Artificial Intelligence company over political and contractual conflict.

Anumana wins FDA clearance for pulmonary hypertension ECG Artificial Intelligence tool

Anumana has received FDA 510(k) clearance for an Artificial Intelligence-enabled pulmonary hypertension algorithm designed for use with standard 12-lead electrocardiograms. The company says the software can help clinicians spot early signs of disease within existing workflows and without moving patient data outside the health system environment.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.