sourc.dev
Home LLMs Tools SaaS APIs
Claude 3.5 Sonnet input $3.00/1M ↓ -50%
GPT-4o input $2.50/1M
Gemini 1.5 Pro input $1.25/1M
Mistral Large input $2.00/1M ↓ -33%
DeepSeek V3 input $0.27/1M
synced 2026-04-05
Claude 3.5 Sonnet input $3.00/1M ↓ -50%
GPT-4o input $2.50/1M
Gemini 1.5 Pro input $1.25/1M
Mistral Large input $2.00/1M ↓ -33%
DeepSeek V3 input $0.27/1M
synced 2026-04-05
#27 of 50

Batch pricing

Some API calls cost half as much if you can wait

What is batch pricing

Batch pricing is a discounted rate offered by model providers for API requests submitted in bulk with no latency guarantee. Instead of processing each request immediately, the provider queues batch requests and processes them when compute capacity is available — typically within 24 hours.

OpenAI offers 50% off for api" class="glossary-link">batch API requests. Anthropic offers a similar discount. The model output is identical — the only difference is response time.

Why it matters

If your workload does not require real-time responses — data processing, content generation at scale, overnight evaluation runs — batch pricing cuts your model cost in half. The constraint is latency, not capability.

Verified March 2026 · Source: OpenAI batch API docs

Related terms
Input priceOutput priceRate limit
← All terms
← Context caching Price per request →