Models with Caching
LLMs with Prompt Caching: Save on Repeated API Calls
Compare LLMs that offer prompt caching. Cache your system prompt, few-shot examples, or conversation history to dramatically cut costs on subsequent requests. Essential for AI agents and chatbots that make many calls with shared context.
Top Picks
Most Capable
Deepseek R1
DeepSeek
$0.7000/1M input
Largest Context
Deepseek Chat
DeepSeek
163.8Kcontext
Compare Models
Related Categories
Coding Models
Top models for code generation, debugging, and software engineering — ranked by LiveBench coding scores and API price.
Writing Models
Models with the highest language and writing scores for content creation, copywriting, marketing, and creative tasks.
Reasoning Models
Models with extended thinking for complex math, logic, and multi-step problem solving — ranked by benchmarks.
Models for AI Agents
Models built for autonomous AI agents — with function calling, large context windows, and competitive pricing for agentic workflows.
Cost per 1M Tokens / Flagship Models
All Models with Caching
| Provider | Model | Input $/1M | Output $/1M | Context | Quality |
|---|---|---|---|---|---|
| DeepSeek | Deepseek V3.2 | $0.2600 | $0.3800 | 163.8K | 84.8Frontier |
| DeepSeek | Deepseek Chat | $0.3200 | $0.8900 | 163.8K | 83.4Frontier |
| DeepSeek | Deepseek R1 | $0.7000 | $2.50 | 64K | 64.6Strong |
Deploy your AI agent for $33/mo flat.
Managed Telegram bot hosting. We handle the infrastructure.
Any model. Any channel. Zero infrastructure.