Google · released 2025-09-01 · text model
Gemini 2.5 Pro
AA INDEX87editorial 2026-05editorial estimate
Goes to OpenRouter · referral link · See disclosure →
Input price$1.25/1M tokens
Output price$10.00/1M tokens
Context1Mtokens
P50 latency0.8sfirst tokvendor-reported
Throughput140tok/svendor-reported
The largest context window in the top tier (1M tokens), which makes it the natural pick for whole-codebase code review, full-document legal analysis, and long video transcript processing. Multimodal handling is best-in-class on image + text. Slower first-token than GPT-5, pricier than DeepSeek V3.5 — choose when context size is the constraint.
Featured in:GPT-5 dominated the AI conversation this week
Full pricing details
all prices in USD per 1M tokens · All hidden costs surfaced| Pricing tier | Input | Output | Notes |
|---|---|---|---|
Standard Pay-as-you-go API rate card | $1.25 | $10.00 | Default if no headers set |
With prompt caching Cached read · cache write | $0.13−90% | $10.00 | Cache writes priced at standard input ($1.25); cached reads 90% off. Requires cache_control header. |
Batch API (50% off) 24h SLA, async | $0.63−50% | $5.00−50% | Send up to 50K requests in one batch; results within 24h |
Image input (vision) Per image, low/high detail differ | $1.25 + image fee | $10.00 | $0.50 per 1024×1024 image at high detail; low detail ~$0.10 |
Reasoning tokens Hidden chain-of-thought tokens | $1.25 | $10.00 × ~3-8 | When using reasoning.effort: high, output tokens include hidden reasoning. Real cost can be 3-8× headline output. |
Structured output JSON mode + schema | $1.25 | $10.00 | No extra cost; first call may be slower (~2x latency) due to schema compilation |
Capabilities
Vision (image understanding)Function callingStructured output (JSON schema)Reasoning modeMultilingual (32+ langs)Batch APIAudio I/OFine-tuning