Skip to content

DeepSeek R1 671B

by DeepSeek · deepseek-r1 family

671B

parameters

text-generation code-generation reasoning math creative-writing summarization

DeepSeek R1 671B is the full-scale reasoning model from DeepSeek, trained with reinforcement learning to perform deep chain-of-thought reasoning. Unlike the distilled variants, this is the original teacher model with 671B mixture-of-experts parameters, delivering state-of-the-art reasoning performance. The model excels at complex mathematical proofs, multi-step logical reasoning, and challenging coding problems by explicitly showing its reasoning process. Its MIT license makes it one of the most permissively licensed frontier-class models available.

Quick Start with Ollama

ollama run 671b-q4_K_M
Resources Ollama Hugging Face Official Page Research Paper
Creator DeepSeek
Parameters 671B
Architecture mixture-of-experts
Context 128K tokens
Released Jan 20, 2025
License MIT
Ollama deepseek-r1:671b

Quantization Options

Format File Size VRAM Required Quality Ollama Tag
Q4_K_M rec 350 GB 362 GB 671b-q4_K_M
Q8_0 670 GB 685 GB 671b-q8_0

Compatible Hardware

Q4_K_M requires 362 GB VRAM

Compatible Hardware

HardwareVRAMTypeFitEst. Speed
Mac Studio M4 Ultra 512GB512 GBmacRuns~2 tok/s
106 hardware device(s) cannot run this model at Q4_K_M.

Benchmark Scores

90.8
mmlu