Sarvam AI Unveils Flagship LLM Surpassing Language and Reasoning Benchmarks

Bengaluru’s Sarvam AI debuts its 24-billion-parameter multilingual model, outperforming global leaders in Indian languages, math, and code for advanced Artificial Intelligence applications.

Bengaluru-based Sarvam AI, a start-up selected under the Indian government´s IndiaAI Mission, has introduced its flagship Large Language Model (LLM) named Sarvam-M. The new model is a 24-billion-parameter, multilingual, hybrid-reasoning, text-only solution built upon Mistral Small—an open-weight model from French company Mistral AI. Sarvam-M is designed for a range of applications including conversational agents, translation tools, and educational services, reflecting its broad ambitions to support India’s unique linguistic and technological landscape.

Sarvam-M sets new performance benchmarks in Indian languages, mathematical reasoning, and programming competence. The company reports that the model delivers a 20% average improvement over its base on Indian language tasks, a 21.6% enhancement in math problems, and a 17.6% uptick on coding benchmarks. Particularly striking is its +86% improvement on romanized Indian language GSM-8K math benchmarks, underscoring its strength in combining linguistic and logical reasoning relevant to local contexts. This release follows the firm’s earlier rollout of Bulbul, a speech model supporting 11 Indian languages and authentic accents.

In comparative evaluations, Sarvam-M is said to outperform Meta’s LLaMA-4 Scout and is competitive with larger models such as LLaMA-3.3 70B and Google’s Gemma 3 27B, even though those models are trained on significantly more data. While Sarvam-M exhibits slightly diminished scores (about 1% below baseline) on English knowledge benchmarks, it is made available as open source on Hugging Face, with API access for developers, promoting adaptability and research. Its architecture uniquely supports two operating modes: ´think´ mode for tasks requiring complex logic and computation, and ´non-think´ mode for efficient, general-purpose text generation and conversation. Importantly, Sarvam-M is post-trained on Indian languages alongside English, ensuring robust Indic cultural representation, and supports both native Indic scripts and romanized forms for diverse user needs.

72

Impact Score

The missing step between Artificial Intelligence hype and profit

Artificial Intelligence companies have built powerful systems and promised sweeping change, but the path from technical progress to real business value remains unclear. Conflicting studies, weak workplace performance, and poor transparency are leaving a critical gap between hype and evidence.

Samsung workers leaked secrets into ChatGPT

Samsung employees reportedly exposed confidential company information while using ChatGPT for coding help and meeting note generation. The incidents highlight the risk of feeding sensitive data into public Artificial Intelligence tools that retain user inputs.

DeepSeek launches new flagship Artificial Intelligence models

DeepSeek has introduced preview versions of its V4 Flash and V4 Pro models, positioning them as its most powerful open-source Artificial Intelligence platform yet. The release renews competition with OpenAI, Anthropic, and major Chinese rivals while drawing fresh attention to the startup’s technical ambitions and regulatory scrutiny.

OpenAI’s GPT-5.5 sharpens coding but trails Anthropic’s Opus 4.7

OpenAI’s latest model upgrade improves coding, tool use, reasoning and token efficiency as the company pushes deeper into enterprise adoption. Early evaluations suggest stronger security performance, but Anthropic’s Opus 4.7 still leads in some important coding areas.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.