Open-weight models

66 models tracked. Hardware requirements, license, and quantization sizes for each.

qwen

Qwen 3 235B-A22B

235B

Qwen 3 flagship MoE. 235B total / 22B active per token, with built-in 'thinking' and 'non-thinking' modes that trade speed for reasoning dep

Commercial OK
128K ctx

Qwen 3 30B-A3B

30B

Mid-tier Qwen 3 MoE. 30B total / 3B active means 70B-class quality at 7B-class inference speed on a single 24GB card. The sweet spot of the

Commercial OK
128K ctx

Qwen 2.5 Coder 32B Instruct

32B

Coding-specialist Qwen 2.5. Beats GPT-4o on HumanEval and matches Sonnet on many code-edit benchmarks. The default local-coding model on 24G

Commercial OK
128K ctx

Qwen 3 32B

32B

Dense Qwen 3 32B. Best dense open-weight model in its size class at release; pairs nicely with a single RTX 5090 or 4090.

Commercial OK
128K ctx

Qwen 3 8B

8B

Qwen 3 at the 8B scale. Direct head-to-head against Llama 3.1 8B on most benchmarks; usually wins on coding and structured output.

Commercial OK
128K ctx

Qwen 3 14B

14B

14B Qwen 3. Fits on 12GB cards at Q4. Strong default for users with a single mid-range GPU.

Commercial OK
128K ctx

Qwen 2.5 7B Instruct

7B

The community-default small Qwen prior to Qwen 3. Still widely used because of mature ecosystem support.

Commercial OK
128K ctx

Qwen 2.5 14B Instruct

14B

14B Qwen 2.5. Sweet spot for 16GB VRAM. Many production deployments still on this version.

Commercial OK
128K ctx

Qwen 2.5 32B Instruct

32B

Dense 32B Qwen 2.5. Strong daily-driver on 24GB cards prior to Qwen 3 32B.

Commercial OK
128K ctx

QwQ 32B Preview

32B

Qwen team's reasoning-focused experimental release. Visible chain-of-thought in <think> tags. Precursor to Qwen 3's thinking mode.

Commercial OK
32K ctx

Qwen 2.5 72B Instruct

72B

The flagship of Qwen 2.5. Workstation-tier; needs 48GB+ VRAM for usable inference.

Commercial OK
128K ctx

Qwen 3 4B

4B

Compact Qwen 3 for edge and laptop deployment. Outperforms many 7B models from prior generations.

Commercial OK
128K ctx

llama

Llama 3.1 8B Instruct

8B

Meta's small flagship. Strong general reasoning, 128K context, broad multilingual. The default first try for most local-AI use cases on cons

Commercial OK
128K ctx

Llama 4 Scout

109B

Meta's 2026 flagship MoE model. 109B total parameters with only 17B active per forward pass and a record 10-million-token context window — u

Commercial OK
9766K ctx

Llama 3.3 70B Instruct

70B

Late-2024 refresh of the 70B Llama line. Roughly matches Llama 3.1 405B on most benchmarks at one-fifth the parameter count. The default hig

Commercial OK
128K ctx

Llama 3.2 3B Instruct

3B

Lightweight 3B for edge and laptop deployment. Runs comfortably on 8GB VRAM at 30+ tok/s on Apple Silicon.

Commercial OK
128K ctx

Llama 3.1 70B Instruct

70B

The 70B sibling of Llama 3.1 8B. Strong generalist reasoning with 128K context, popular base for agentic fine-tunes (Hermes 3, Nemotron). Mo

Commercial OK
128K ctx

Llama 3.1 Nemotron 70B Instruct

70B

NVIDIA's HelpSteer2-tuned Llama 3.1 70B. Topped Arena Hard at release. The pre-Nemotron-3 NVIDIA reference open weights.

Commercial OK
128K ctx

Llama 3.2 11B Vision Instruct

11B

First-party multimodal Llama. Accepts images alongside text for VQA, document understanding, and chart reading. Runs on 12GB+ VRAM.

Commercial OK
Multimodal
128K ctx

Llama 4 Maverick

400B

Meta's high-end Llama 4 sibling — 128 experts MoE built for performance over efficiency. Multilingual strength is its standout. Effectively

Commercial OK
Multimodal
977K ctx

Llama 3.1 Nemotron Ultra 253B

253B

NVIDIA's top open reasoning model in the Llama 3.1 lineage. Server-tier; trained for groundbreaking reasoning accuracy on agentic workloads.

Commercial OK
128K ctx

Llama 3.1 Nemotron Nano 8B

8B

Smallest of the Nemotron reasoning trio. NAS-optimized for inference efficiency on RTX hardware.

Commercial OK
128K ctx

Llama 3.2 1B Instruct

1B

True edge-tier Llama. Runs on a phone or Raspberry Pi. Useful for classification, simple summarization, and on-device agents.

Commercial OK
128K ctx

Llama 3.2 90B Vision Instruct

90B

The 90B vision Llama. Best-in-class first-party multimodal open weight at the time of release. Workstation-class only.

Commercial OK
Multimodal
128K ctx

deepseek

gemma

other

mistral

phi

hermes

dolphin

mixtral

command-r

yi

wizard