DeepSeek R1 671B
by DeepSeek · deepseek-r1 family
671B
parameters
text-generation code-generation reasoning math creative-writing summarization
DeepSeek R1 671B is the full-scale reasoning model from DeepSeek, trained with reinforcement learning to perform deep chain-of-thought reasoning. Unlike the distilled variants, this is the original teacher model with 671B mixture-of-experts parameters, delivering state-of-the-art reasoning performance. The model excels at complex mathematical proofs, multi-step logical reasoning, and challenging coding problems by explicitly showing its reasoning process. Its MIT license makes it one of the most permissively licensed frontier-class models available.
Quick Start with Ollama
ollama run 671b-q4_K_M | Creator | DeepSeek |
| Parameters | 671B |
| Architecture | mixture-of-experts |
| Context | 128K tokens |
| Released | Jan 20, 2025 |
| License | MIT |
| Ollama | deepseek-r1:671b |
Quantization Options
| Format | File Size | VRAM Required | Quality | Ollama Tag |
|---|---|---|---|---|
| Q4_K_M rec | 350 GB | 362 GB | | 671b-q4_K_M |
| Q8_0 | 670 GB | 685 GB | | 671b-q8_0 |
Compatible Hardware
Q4_K_M requires 362 GB VRAM
Benchmark Scores
90.8
mmlu