Anthropic · released 2026-04-08 · text model
Claude Haiku 4.5
AA INDEX79editorial 2026-05editorial estimate
Goes to OpenRouter · referral link · See disclosure →
Input price$1.00/1M tokens
Output price$5.00/1M tokens
Context200Ktokens
P50 latency0.4sfirst tokvendor-reported
Throughput250tok/svendor-reported
Anthropic's economy tier. Inherits the family's writing voice and safer refusal patterns at a price point closer to GPT-5-mini. Strong choice when output voice or the audience's regulatory expectations matter, but raw reasoning headroom is below the frontier — keep it for chat-shaped workloads rather than agentic chains.
Featured in:The Cheapest LLM API for Production in 2026
Full pricing details
all prices in USD per 1M tokens · All hidden costs surfaced| Pricing tier | Input | Output | Notes |
|---|---|---|---|
Standard Pay-as-you-go API rate card | $1.00 | $5.00 | Default if no headers set |
With prompt caching Cached read · cache write | $0.10−90% | $5.00 | Cache writes priced at standard input ($1.00); cached reads 90% off. Requires cache_control header. |
Batch API (50% off) 24h SLA, async | $0.50−50% | $2.50−50% | Send up to 50K requests in one batch; results within 24h |
Image input (vision) Per image, low/high detail differ | $1.00 + image fee | $5.00 | $0.50 per 1024×1024 image at high detail; low detail ~$0.10 |
Reasoning tokens Hidden chain-of-thought tokens | $1.00 | $5.00 × ~3-8 | When using reasoning.effort: high, output tokens include hidden reasoning. Real cost can be 3-8× headline output. |
Structured output JSON mode + schema | $1.00 | $5.00 | No extra cost; first call may be slower (~2x latency) due to schema compilation |
Capabilities
Vision (image understanding)Function callingStructured output (JSON schema)Reasoning modeMultilingual (32+ langs)Batch APIAudio I/OFine-tuning