Llama 4 Maverick
Meta · released 2026-01-25
Quality (AA Index)80
Input price$0.50
Output price$1.50
Context256K
Throughput220 tok/s
P50 latency0.5s
Side-by-side pricing, capabilities, real-world cost across common scenarios, and our editorial pick.
Meta · released 2026-01-25
DeepSeek · released 2026-02-15
Links open in a new tab via our OpenRouter referral. Affiliate disclosure
| Metric | Llama 4 Maverick | DeepSeek R2 | Verdict |
|---|---|---|---|
Input price /1M tokens | $0.50 | $0.55 | Llama 4 Maverick −9% |
Output price /1M tokens | $1.50 | $2.20 | Llama 4 Maverick −32% |
Context window max input length | 256K | 128K | Llama 4 Maverick +2.0× |
AA Quality AA Intelligence Index (0–100) | 80 | 86 | DeepSeek R2 +6pt |
Arena Elo LMArena human-pref Elo (800–2000) | — | — | Tied |
Throughput tokens per second | 220 | 110 | Llama 4 Maverick +100% |
P50 latency first token | 0.5s | 1.2s | Llama 4 Maverick −58% |
Vision multimodal | — | ||
Function calling tool use | Tied | ||
Reasoning mode chain-of-thought | — |
| Scenario | Llama 4 Maverick | DeepSeek R2 |
|---|---|---|
customer support 1000K req · 600/180 tok | $570 | $726 |
chat with docs 300K req · 4000/300 tok | $735 | $858 |
code generation 500K req · 2000/500 tok | $875 | $1.1K |
voice assistant 600K req · 800/200 tok | $420 | $528 |
Best open-weight alternative for self-hosted deployments. Quality lags closed-source frontier but cost-per-token via inference providers is the lowest in the table.
Set Llama 4 Maverick as primary, DeepSeek R2 as fallback. One key, one bill, automatic failover when Llama 4 Maverick errors.
$ curl https://api.aipricly.com/v1/chat/completions \
-H "Authorization: Bearer $AIPC_KEY" \
-d '{
"routing": {
"primary": "meta/llama-4-maverick",
"fallback": ["deepseek/deepseek-r2"]
},
"messages": [{"role": "user", "content": "..."}]
}'