Mistral Open-Source Models
Mistral AI (Paris) — From Mistral 7B (Sep. 2023) to Mistral Large 3 (Dec. 2025)
At a glance: Mistral AI publishes Europe's most comprehensive open-weight model family, from the tiny Ministral 3B (runnable on a smartphone) to the massive Mistral Large 3 (675 billion parameters). All under Apache 2.0 license — free commercial use, modification, and redistribution. For clinicians, this is the most relevant open-weight ecosystem: European, natively French-speaking, with models at every scale. Combined with local hosting, it offers the most complete data sovereignty available today.
Model catalog (December 2025)
Mistral offers models at every scale, from the lightest (3B parameters, runnable on a phone) to the most powerful (675B, requiring a data center). Here are the most relevant models for healthcare use.
Small models (edge / workstation)
| Model | Size | Date | VRAM (Q4) |
|---|---|---|---|
| Ministral 3B | 3B | Dec. 2025 | ~2-3 GB |
| Mistral 7B | 7.3B | Sep. 2023 | ~5-6 GB |
| Ministral 8B | 8B | Dec. 2025 | ~5-6 GB |
| Ministral 14B | 14B | Dec. 2025 | ~10 GB |
Medium models (dedicated server)
| Model | Size | Date | VRAM (Q4) |
|---|---|---|---|
| Mistral Nemo | 12B | Jul. 2024 | ~8 GB |
| Mistral Small 3.1 | 24B | Mar. 2025 | ~16 GB |
| Mixtral 8x7B | 47B (13B active) | Jan. 2024 | ~26 GB |
| Mixtral 8x22B | 141B (39B active) | Apr. 2024 | ~80 GB |
Frontier model
| Model | Size | Date | Infrastructure |
|---|---|---|---|
| Mistral Large 3 | 675B (41B active) | Dec. 2025 | Data center required |
Specialized models (Apache 2.0)
- • Pixtral 12B: multimodal (text + image)
- • Codestral Mamba 7B: code generation, Mamba2 architecture
- • Devstral Small 2 (24B): advanced coding agent
- • Magistral Small (24B): reasoning (chain-of-thought)
Local installation: practical guide
The major advantage of open-weight models is being able to run them on your own hardware. No data leaves your infrastructure. Here are recommendations based on your equipment.
By hardware
- • Mac M1/M2 (8-16 GB): Ministral 3B, Mistral 7B quantized
- • RTX 4060/Mac M2 Pro (16 GB): Mistral 7B, Ministral 8B
- • RTX 4090/Mac M2 Max (24-32 GB): Ministral 14B, Mistral Small 3.1 quantized
- • Multi-GPU/Mac Studio (48+ GB): Mistral Small 3.1 full precision, Mixtral 8x7B
- • Dedicated server (64+ GB): Mixtral 8x22B quantized
Recommended tools
- • Ollama: The simplest. A single command:
ollama run mistral. Automatically handles downloading and quantization. - • LM Studio: Graphical interface, ideal for beginners. Direct download from Hugging Face.
- • llama.cpp: More control, supports Apple Silicon, NVIDIA, AMD. Ollama's underlying engine.
- • vLLM: Optimized for serving multiple users (institutional settings).
Technical note: MoE models (Mixtral, Large 3) load all experts into memory, not just the active ones. Mixtral 8x7B (47B parameters) therefore requires the same VRAM as a dense 47B model — not 13B. This is a common pitfall. To understand MoE architecture, see our glossary entry on LLMs.
Mental health scenarios
Healthcare facility
- • Sovereign clinical writing assistance: Mistral Small 3.1 on an internal server for drafting reports without data leakage.
- • Research and training: Mistral Large 3 via API for literature review, combined with a local model for processing sensitive data.
- • Health data hosting compliance: Deployment on a certified health data hosting provider for full compliance.
Private practitioner
- • Ministral 8B on a Mac: Lightweight model for exploring clinical hypotheses locally, without sending data externally.
- • Psychoeducation: Generate educational materials (worksheets, metaphors) with a local model.
Research
- • BioMistral: Community model based on Mistral 7B, pre-trained on PubMed Central (CNRS Jean Zay). Strictly for research — not for production clinical use.
- • Fine-tuning: The Apache 2.0 license allows specialized training on French clinical corpora — an impossibility with proprietary models.
The open-weight / open-source nuance
As with GPT-OSS, Mistral's models are more precisely open-weight than open-source in the strict sense: the trained weights and inference code are public, but the training data and alignment procedures are not.
For clinicians, this means: you control where the model runs and who accesses the data, but you cannot verify what it was trained on. Mistral's European advantage (francophone culture, French team) makes it plausible that French and European contexts are better represented in the training data — but this is an inference, not a certainty.
Limitations and precautions
Technical skills required
Local installation requires minimum technical skills. Ollama greatly simplifies the process, but managing an institutional deployment remains an IT project.
Hallucinations
A local model hallucinates just as much as a cloud model. Sovereignty applies to data, not to reliability.
No built-in guardrails
Open-weight models inherit basic alignment, but advanced safety mechanisms (crisis detection, emergency redirection) must be implemented by the organization.
Variable quality
Small models (3B, 7B) are significantly less capable than frontier models. For tasks requiring nuance and reasoning (clinical analysis), prefer Mistral Small 3.1 or above.
Our analysis
Mistral's open-weight ecosystem is the most comprehensive available in Europe. It combines three advantages that neither OpenAI (GPT-OSS) nor Meta (LLaMA) offer simultaneously: a European company, native multilingualism centered on French, and models at every scale.
For a French mental health facility, the most realistic short-term scenario is: Mistral Small 3.1 on an internal server (or certified health data host) for writing assistance and clinical exploration, complemented by Le Chat Enterprise for uses requiring the frontier model.
The release of Mistral Large 3 under Apache 2.0 (December 2025) is a strong strategic signal: for the first time, a European frontier model of 675 billion parameters is freely downloadable and usable by any institution. Even though running it requires a data center, this opens the way to sovereign deployments at the scale of a hospital group (GHT — Groupement Hospitalier de Territoire) or a regional health agency (ARS — Agence Régionale de Santé).
Related concepts and terms
A risk independent of hosting location
Even locally, informing remains necessary
Can sovereignty reduce distrust?
Protecting data as an act of care
Glossary entries
See also
References
Mistral AI (2025). Introducing Mistral 3. mistral.ai/news/mistral-3.
Labrak, Y. et al. (2024). BioMistral: A Collection of Open-Source Pretrained LLMs for Medical Domains. arXiv:2402.10373.
HAS (2025). First guidelines for generative AI use in healthcare — A.V.E.C. Framework.
Model card last updated: February 2026