Model Comparison

Baseline scores + community votes · 37 models

Provider
Tier
Sort by
Cost estimate for:tokens · click row to expand scores
ModelRadarValue ↕OverallComplexReasoningCodingSpeedEst. CostInput/1M
DeepSeek V3★ Best value
DeepSeek
balanced
7.6
7.9
8.0
8.0
8.5
7.0
$0.049$0.270
Llama 4 Maverick
Together
balanced
7.5
7.8
7.5
7.5
7.5
8.0
$0.032$0.270
Gemini 2.0 Flash
Google
fast
7.4
8.0
7.0
7.0
7.5
9.5
$0.018$0.100
Codestral
Mistral
balanced
7.4
7.9
7.0
7.0
9.5
8.0
$0.048$0.300
GLM-4 Plus
Zhipu
flagship
7.4
7.6
7.5
7.5
7.0
8.0
$0.017$0.140
Qwen Max
Alibaba
flagship
7.3
7.7
8.0
8.0
7.5
7.0
$0.064$0.400
Qwen Plus
Alibaba
balanced
7.3
7.5
7.0
7.0
7.0
8.0
$0.013$0.080
Llama 3.3 70B
Groq
balanced
7.2
8.2
7.5
7.5
7.5
9.8
$0.075$0.590
Grok 3 Mini
xAI
fast
7.2
7.7
6.5
8.0
7.0
8.5
$0.040$0.300
MiniMax Text 01
MiniMax
flagship
7.0
7.5
7.5
7.5
7.0
7.5
$0.042$0.200
GPT-4o Mini
OpenAI
fast
7.0
7.6
6.5
6.5
7.0
9.0
$0.027$0.150
DeepSeek R1
DeepSeek
flagship
6.9
7.0
9.0
9.5
8.5
3.5
$0.099$0.550
Moonshot v1 128K
Moonshot
balanced
6.8
7.4
7.5
7.0
7.0
7.5
$0.098$0.820
Moonshot v1 8K
Moonshot
fast
6.8
7.3
6.0
6.5
6.5
9.0
$0.014$0.120
DeepSeek R1 Distill 70B
Groq
balanced
6.8
7.7
8.0
9.0
7.5
9.0
$0.095$0.750
Llama 3.3 70B Turbo
Together
balanced
6.7
7.5
7.0
7.0
7.0
8.5
$0.106$0.880
Gemini 2.0 Flash Lite
Google
fast
6.6
7.4
5.5
5.5
6.5
9.8
$0.013$0.075
ABAB 6.5s
MiniMax
fast
6.6
7.1
6.0
6.0
6.0
9.0
$0.012$0.100
Qwen Turbo
Alibaba
fast
6.5
7.1
5.5
5.5
6.0
9.5
$0.0032$0.020
Mistral Small
Mistral
fast
6.3
6.9
5.5
5.5
6.0
9.0
$0.016$0.100
Sonar
Perplexity
fast
6.2
7.2
6.5
6.5
6.0
8.5
$0.120$1.00
Command R
Cohere
balanced
6.1
6.7
6.0
6.0
6.0
8.0
$0.027$0.150
GLM-4 Flash
Zhipu
fast
6.1
6.8
5.0
5.5
5.5
9.5
FreeFree
Llama 3.1 8B Instant
Groq
fast
6.1
6.9
5.0
5.5
5.5
10.0
$0.0066$0.050
o4-mini
OpenAI
balanced
6.1
7.3
8.0
9.0
8.0
5.5
$0.198$1.10
Jamba 1.5 Mini
AI21
fast
5.9
6.7
5.5
5.5
5.5
9.0
$0.028$0.200
Claude Haiku 4.5
Anthropic
fast
5.7
7.6
6.0
6.0
7.0
9.5
$0.160$0.800
Gemini 2.5 Pro
Google
flagship
5.5
8.2
9.0
9.0
8.5
6.5
$0.325$1.25
Mistral Large
Mistral
flagship
5.4
7.7
7.5
7.5
7.5
8.0
$0.320$2.00
GPT-4o
OpenAI
flagship
5.2
8.3
8.5
8.0
8.5
8.0
$0.450$2.50
Jamba 1.5 Large
AI21
flagship
4.8
7.3
7.5
7.0
7.0
7.5
$0.360$2.00
Claude Sonnet 4.5
Anthropic
balanced
4.5
8.2
8.5
8.5
8.5
7.0
$0.600$3.00
Grok 3
xAI
flagship
4.5
8.2
8.5
9.0
8.5
7.0
$0.600$3.00
Command R+
Cohere
flagship
4.5
7.3
7.5
7.0
7.0
7.5
$0.450$2.50
Sonar Pro
Perplexity
flagship
4.1
7.6
8.0
7.5
7.0
7.0
$0.600$3.00
o3
OpenAI
flagship
2.3
7.2
9.5
9.8
9.0
3.0
$1.80$10.00
Claude Opus 4.5
Anthropic
flagship
1.6
7.9
9.5
9.5
9.0
4.5
$3.00$15.00
↑↓ Community votes shift scores ±0.1 per net vote·Value = quality ÷ (price × 0.1 + 1)·Click row to expand