deepseek
671B parameters
Commercial OK

DeepSeek R1 (671B reasoning)

Open reasoning model that closed the gap with frontier proprietary reasoners. Visible chain-of-thought, MIT license, and a family of distilled smaller variants.

License: MIT·Released Jan 20, 2025·Context: 131,072 tokens
Our verdict
By Fredoline Eruo·Last verified May 6, 2026
9.0/10
Positioning

DeepSeek R1 is the o1-equivalent open-weight model — explicit reasoning training, visible chain-of-thought, state-of-the-art on math and competitive programming benchmarks. Same MoE architecture as V3, same workstation-class hardware requirement.

Strengths
  • Reasoning ceiling matches closed frontier models — true o1-class on hard math and code planning.
  • Fully open weights — uniquely valuable in the reasoning space where most leaders are closed.
  • Clean MIT-style license.
Limitations
  • Workstation hardware required — same ~380 GB footprint as V3.
  • Verbose chain-of-thought consumes lots of tokens.
  • Distill versions exist (R1 Distill 70B, 32B, 14B, 7B) — those are the practical local picks.
Real-world performance on RTX 4090
  • Direct R1 Q4_K_M (~380 GB) — workstation only, same as V3
  • Practical local path: run R1 Distill Llama 70B or R1 Distill Qwen 32B (much more accessible)
Should you run this locally?

Yes, for workstation owners — same hardware story as V3. No, for consumer hardware — pick the R1 Distill variants instead, which deliver most of the reasoning quality at viable hardware costs.

How it compares
  • vs DeepSeek V3 → R1 is the reasoning specialist, V3 is the generalist. Different jobs.
  • vs DeepSeek R1 Distill Llama 70B → Distill is much more accessible (single 4090 with offload) and captures most of the reasoning lift. Default pick for local hardware.
  • vs QwQ 32B → QwQ is the reasoning specialist that fits on a single 4090; R1 has higher ceiling.
  • vs OpenAI o1 → R1 is the open-weight equivalent; quality competitive on math/code.
Run this yourself
# For local hardware, prefer the distills:
ollama pull deepseek-r1:70b-distill-llama-q4_K_M
ollama pull deepseek-r1:32b-distill-qwen-q4_K_M
Direct R1 settings: Q4_K_M, multi-GPU, A100/H100 cluster
Why this rating

9.0/10 — DeepSeek's reasoning specialist matches o1-class performance on hard problems and is fully open-weight. Same workstation-size reality as V3. Loses fractional points only on hardware barrier.

Overview

Open reasoning model that closed the gap with frontier proprietary reasoners. Visible chain-of-thought, MIT license, and a family of distilled smaller variants.

Strengths

  • MIT license
  • Frontier reasoning quality
  • Visible CoT

Weaknesses

  • 671B is server-only
  • Verbose by default

Quantization variants

Each quantization trades model quality for file size and VRAM. Q4_K_M is the most popular starting point.

QuantizationFile sizeVRAM required
Q4_K_M380.0 GB420 GB

Get the model

Ollama

One-line install

ollama run deepseek-r1:671bRead our Ollama review →

HuggingFace

Original weights

huggingface.co/deepseek-ai/DeepSeek-R1

Source repository — direct quantization required.

Hardware that runs this

Cards with enough VRAM for at least one quantization of DeepSeek R1 (671B reasoning).

Compare alternatives

Models worth comparing

Same parameter band, plus what's one tier above and below — so you can decide what actually fits your hardware.

Step up
More capable — bigger memory footprint
No verdicted models in the next tier up yet.

Frequently asked

What's the minimum VRAM to run DeepSeek R1 (671B reasoning)?

420GB of VRAM is enough to run DeepSeek R1 (671B reasoning) at the Q4_K_M quantization (file size 380.0 GB). Higher-quality quantizations need more.

Can I use DeepSeek R1 (671B reasoning) commercially?

Yes — DeepSeek R1 (671B reasoning) ships under the MIT, which permits commercial use. Always read the license text before deployment.

What's the context length of DeepSeek R1 (671B reasoning)?

DeepSeek R1 (671B reasoning) supports a context window of 131,072 tokens (about 131K).

How do I install DeepSeek R1 (671B reasoning) with Ollama?

Run `ollama pull deepseek-r1:671b` to download, then `ollama run deepseek-r1:671b` to start a chat session. The default quantization is Q4_K_M.

Source: huggingface.co/deepseek-ai/DeepSeek-R1

Reviewed by RunLocalAI Editorial. See our editorial policy for how we research and verify model claims.