DeepSeek R1

DeepSeek · China
Frontier reasoning model at o1-class performance. MIT licence makes weights legally clean. Same Chinese-origin alignment/supply-chain considerations as DeepSeek V3. Distilled Qwen/Llama versions inherit their base licence.
Licence facts
Licence
MIT
Commercial use
Unrestricted
Derivatives
Allowed (including distillation)
Attribution
Minimal
Parameters
671B MoE (37B active)
Context
128K
Training data
Not disclosed
Last updated
2025-Q1
Known risks
  • Chinese-origin alignment biases (political, historical)
  • Self-hosting 671B MoE is expensive even quantized
  • Reasoning traces may leak chain-of-thought considered sensitive in some workflows
Reviewed by Ali Madjaji · Last reviewed 2026-04-15