NVIDIA partners with Mistral Artificial Intelligence on open-source Mistral 3 models

NVIDIA and Mistral Artificial Intelligence launched the open-source Mistral 3 family, a suite of multilingual, multimodal models engineered for NVIDIA hardware. The release pairs mixture-of-experts efficiency with optimizations for GB200 NVL72 systems and edge platforms.

NVIDIA and Mistral Artificial Intelligence unveiled the Mistral 3 family, a set of open-source multilingual, multimodal models built and optimized for NVIDIA’s supercomputing and edge platforms. The companies position the release as an enterprise-focused offering that brings mixture-of-experts model architecture to production workloads and edge devices. According to the announcement, the partnership aims to bridge research breakthroughs with real-world applications and to advance what they call the era of distributed intelligence.

The flagship Mistral Large 3 uses a mixture-of-experts approach that activates only the most relevant parts of the model for each query, so it does not run all parameters for every task. The model cites “41B active parameters” and “675B total parameters” as part of that design and supports a “256K context window” for long-form inputs. Benchmarks in the article show significant performance improvements on GB200 NVL72 systems compared with the H200 generation, driven by NVLink coherent memory and wide expert parallelism optimizations that raise throughput, lower per-token costs, and improve energy efficiency.

The release also includes nine compact “Ministral 3” models targeted at edge deployment and optimized for NVIDIA Spark, RTX PCs, laptops, and Jetson devices. Developers can access compact variants through frameworks such as Llama.cpp and Ollama. The open-source orientation is intended to contrast with proprietary models from other vendors and to allow enterprises to customize models using tools named in the article, including Data Designer, Customizer, Guardrails, and the NeMo Agent Toolkit. NVIDIA has optimized multiple inference frameworks for the Mistral 3 family, including TensorRT-LLM, SGLang, and vLLM, and says the models are available on leading cloud platforms and open-source repositories with NVIDIA NIM microservices deployment coming soon.

The article frames the partnership as a democratizing move for enterprise Artificial Intelligence, offering high-performance open-source alternatives and enterprise-grade tooling. It highlights immediate developer access to state-of-the-art multimodal capabilities while noting the true test will be adoption by enterprises that need production-ready, scalable, and customizable models without closed-source constraints.

72

Impact Score

Hyperscalers accelerate custom semiconductor and artificial intelligence infrastructure deals in early 2026

Hyperscale cloud providers are ramping multi-gigawatt semiconductor deals across GPUs, custom accelerators, and optical interconnects, with Meta, Google, OpenAI, and Anthropic locking in long-term capacity. Broadcom, AMD, NVIDIA, Marvell, Intel, and MediaTek are reshaping data center and networking roadmaps around custom artificial intelligence silicon and rack-scale systems.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.