Mistral 7B Instruct v0.2
Mistral AI · France
Based on published licence terms, Mistral 7B Instruct v0.2 is an EU-origin open-weight model under standard Apache 2.0 — commercial deployment and self-hosting are permitted without field-of-use restrictions. Training-data opacity is the primary EU AI Act Art. 53 gap, but the French controller, absence of CLOUD Act exposure, and Mistral's published DPA make this a strong baseline for regulated EU workloads.
Licence facts
- Parameters
- ~7.3B dense
- Architecture
- Decoder-only transformer with Grouped-Query Attention and RoPE (theta 1e6); full attention (sliding-window removed vs v0.1)
- Context length
- 32K tokens
- Tensor type
- BF16
- Chat template
- [INST] / [/INST]
- Released
- 2023-12
Known risks
- Training-data transparency gap: Mistral's help centre explicitly states training datasets are not disclosed; the model card and paper describe only 'open web' sources with no corpus breakdown, filtering description, or opt-out mechanism — the primary Art. 53(1)(d) compliance gap, inherited by downstream deployers.
- No built-in safety guardrails: the model card notes the absence of moderation; consumer-facing EU deployments must add their own filtering and downstream-provider documentation under the AI Act.
- Fine-tune provenance is undisclosed: the instruct-tuning dataset is not published, which may matter for sectoral obligations (e.g. DSA Art. 34 risk assessments) or for copyright-TDM opt-out audits.
Sources
See also
Reviewed by Ali Madjaji · Last reviewed 2026-04-17· Reviewed 1 day agoSuggest a correction