DeepSeek R1 (671B reasoning)
Open reasoning model that closed the gap with frontier proprietary reasoners. Visible chain-of-thought, MIT license, and a family of distilled smaller variants.
DeepSeek R1 is the o1-equivalent open-weight model — explicit reasoning training, visible chain-of-thought, state-of-the-art on math and competitive programming benchmarks. Same MoE architecture as V3, same workstation-class hardware requirement.
Strengths- Reasoning ceiling matches closed frontier models — true o1-class on hard math and code planning.
- Fully open weights — uniquely valuable in the reasoning space where most leaders are closed.
- Clean MIT-style license.
- Workstation hardware required — same ~380 GB footprint as V3.
- Verbose chain-of-thought consumes lots of tokens.
- Distill versions exist (R1 Distill 70B, 32B, 14B, 7B) — those are the practical local picks.
- Direct R1 Q4_K_M (~380 GB) — workstation only, same as V3
- Practical local path: run R1 Distill Llama 70B or R1 Distill Qwen 32B (much more accessible)
Yes, for workstation owners — same hardware story as V3. No, for consumer hardware — pick the R1 Distill variants instead, which deliver most of the reasoning quality at viable hardware costs.
How it compares- vs DeepSeek V3 → R1 is the reasoning specialist, V3 is the generalist. Different jobs.
- vs DeepSeek R1 Distill Llama 70B → Distill is much more accessible (single 4090 with offload) and captures most of the reasoning lift. Default pick for local hardware.
- vs QwQ 32B → QwQ is the reasoning specialist that fits on a single 4090; R1 has higher ceiling.
- vs OpenAI o1 → R1 is the open-weight equivalent; quality competitive on math/code.
# For local hardware, prefer the distills:
ollama pull deepseek-r1:70b-distill-llama-q4_K_M
ollama pull deepseek-r1:32b-distill-qwen-q4_K_M
Direct R1 settings: Q4_K_M, multi-GPU, A100/H100 cluster
›Why this rating
9.0/10 — DeepSeek's reasoning specialist matches o1-class performance on hard problems and is fully open-weight. Same workstation-size reality as V3. Loses fractional points only on hardware barrier.
Overview
Open reasoning model that closed the gap with frontier proprietary reasoners. Visible chain-of-thought, MIT license, and a family of distilled smaller variants.
Strengths
- MIT license
- Frontier reasoning quality
- Visible CoT
Weaknesses
- 671B is server-only
- Verbose by default
Quantization variants
Each quantization trades model quality for file size and VRAM. Q4_K_M is the most popular starting point.
| Quantization | File size | VRAM required |
|---|---|---|
| Q4_K_M | 380.0 GB | 420 GB |
Get the model
Ollama
One-line install
ollama run deepseek-r1:671bRead our Ollama review →HuggingFace
Original weights
Source repository — direct quantization required.
Hardware that runs this
Cards with enough VRAM for at least one quantization of DeepSeek R1 (671B reasoning).
Models worth comparing
Same parameter band, plus what's one tier above and below — so you can decide what actually fits your hardware.
Frequently asked
What's the minimum VRAM to run DeepSeek R1 (671B reasoning)?
Can I use DeepSeek R1 (671B reasoning) commercially?
What's the context length of DeepSeek R1 (671B reasoning)?
How do I install DeepSeek R1 (671B reasoning) with Ollama?
Source: huggingface.co/deepseek-ai/DeepSeek-R1
Reviewed by RunLocalAI Editorial. See our editorial policy for how we research and verify model claims.