Open-weight European Apache 2.0

Mistral Open-Source Models

Mistral AI (Paris) — From Mistral 7B (Sep. 2023) to Mistral Large 3 (Dec. 2025)

At a glance: Mistral AI publishes Europe's most comprehensive open-weight model family, from the tiny Ministral 3B (runnable on a smartphone) to the massive Mistral Large 3 (675 billion parameters). All under Apache 2.0 license — free commercial use, modification, and redistribution. For clinicians, this is the most relevant open-weight ecosystem: European, natively French-speaking, with models at every scale. Combined with local hosting, it offers the most complete data sovereignty available today.

Model catalog (December 2025)

Mistral offers models at every scale, from the lightest (3B parameters, runnable on a phone) to the most powerful (675B, requiring a data center). Here are the most relevant models for healthcare use.

Small models (edge / workstation)

Model Size Date VRAM (Q4)
Ministral 3B 3B Dec. 2025 ~2-3 GB
Mistral 7B 7.3B Sep. 2023 ~5-6 GB
Ministral 8B 8B Dec. 2025 ~5-6 GB
Ministral 14B 14B Dec. 2025 ~10 GB

Medium models (dedicated server)

Model Size Date VRAM (Q4)
Mistral Nemo 12B Jul. 2024 ~8 GB
Mistral Small 3.1 24B Mar. 2025 ~16 GB
Mixtral 8x7B 47B (13B active) Jan. 2024 ~26 GB
Mixtral 8x22B 141B (39B active) Apr. 2024 ~80 GB

Frontier model

Model Size Date Infrastructure
Mistral Large 3 675B (41B active) Dec. 2025 Data center required

Specialized models (Apache 2.0)

  • Pixtral 12B: multimodal (text + image)
  • Codestral Mamba 7B: code generation, Mamba2 architecture
  • Devstral Small 2 (24B): advanced coding agent
  • Magistral Small (24B): reasoning (chain-of-thought)

Local installation: practical guide

The major advantage of open-weight models is being able to run them on your own hardware. No data leaves your infrastructure. Here are recommendations based on your equipment.

By hardware

  • Mac M1/M2 (8-16 GB): Ministral 3B, Mistral 7B quantized
  • RTX 4060/Mac M2 Pro (16 GB): Mistral 7B, Ministral 8B
  • RTX 4090/Mac M2 Max (24-32 GB): Ministral 14B, Mistral Small 3.1 quantized
  • Multi-GPU/Mac Studio (48+ GB): Mistral Small 3.1 full precision, Mixtral 8x7B
  • Dedicated server (64+ GB): Mixtral 8x22B quantized

Recommended tools

  • Ollama: The simplest. A single command: ollama run mistral. Automatically handles downloading and quantization.
  • LM Studio: Graphical interface, ideal for beginners. Direct download from Hugging Face.
  • llama.cpp: More control, supports Apple Silicon, NVIDIA, AMD. Ollama's underlying engine.
  • vLLM: Optimized for serving multiple users (institutional settings).

Technical note: MoE models (Mixtral, Large 3) load all experts into memory, not just the active ones. Mixtral 8x7B (47B parameters) therefore requires the same VRAM as a dense 47B model — not 13B. This is a common pitfall. To understand MoE architecture, see our glossary entry on LLMs.

Mental health scenarios

Healthcare facility

  • Sovereign clinical writing assistance: Mistral Small 3.1 on an internal server for drafting reports without data leakage.
  • Research and training: Mistral Large 3 via API for literature review, combined with a local model for processing sensitive data.
  • Health data hosting compliance: Deployment on a certified health data hosting provider for full compliance.

Private practitioner

  • Ministral 8B on a Mac: Lightweight model for exploring clinical hypotheses locally, without sending data externally.
  • Psychoeducation: Generate educational materials (worksheets, metaphors) with a local model.

Research

  • BioMistral: Community model based on Mistral 7B, pre-trained on PubMed Central (CNRS Jean Zay). Strictly for research — not for production clinical use.
  • Fine-tuning: The Apache 2.0 license allows specialized training on French clinical corpora — an impossibility with proprietary models.

The open-weight / open-source nuance

As with GPT-OSS, Mistral's models are more precisely open-weight than open-source in the strict sense: the trained weights and inference code are public, but the training data and alignment procedures are not.

For clinicians, this means: you control where the model runs and who accesses the data, but you cannot verify what it was trained on. Mistral's European advantage (francophone culture, French team) makes it plausible that French and European contexts are better represented in the training data — but this is an inference, not a certainty.

Limitations and precautions

Technical skills required

Local installation requires minimum technical skills. Ollama greatly simplifies the process, but managing an institutional deployment remains an IT project.

Hallucinations

A local model hallucinates just as much as a cloud model. Sovereignty applies to data, not to reliability.

No built-in guardrails

Open-weight models inherit basic alignment, but advanced safety mechanisms (crisis detection, emergency redirection) must be implemented by the organization.

Variable quality

Small models (3B, 7B) are significantly less capable than frontier models. For tasks requiring nuance and reasoning (clinical analysis), prefer Mistral Small 3.1 or above.

Our analysis

Mistral's open-weight ecosystem is the most comprehensive available in Europe. It combines three advantages that neither OpenAI (GPT-OSS) nor Meta (LLaMA) offer simultaneously: a European company, native multilingualism centered on French, and models at every scale.

For a French mental health facility, the most realistic short-term scenario is: Mistral Small 3.1 on an internal server (or certified health data host) for writing assistance and clinical exploration, complemented by Le Chat Enterprise for uses requiring the frontier model.

The release of Mistral Large 3 under Apache 2.0 (December 2025) is a strong strategic signal: for the first time, a European frontier model of 675 billion parameters is freely downloadable and usable by any institution. Even though running it requires a data center, this opens the way to sovereign deployments at the scale of a hospital group (GHT — Groupement Hospitalier de Territoire) or a regional health agency (ARS — Agence Régionale de Santé).

References

Mistral AI (2025). Introducing Mistral 3. mistral.ai/news/mistral-3.

Labrak, Y. et al. (2024). BioMistral: A Collection of Open-Source Pretrained LLMs for Medical Domains. arXiv:2402.10373.

HAS (2025). First guidelines for generative AI use in healthcare — A.V.E.C. Framework.

Model card last updated: February 2026

Back to AI Tools