Pricing & Cost

Cached Tokens

Quick Answer

Previously processed tokens stored and reused, charged at lower rates (prompt caching).

Prompt caching stores previously processed tokens, reusing KV cache. Cached tokens cost 10-25% of standard token price. Caching is valuable when using common context repeatedly (RAG, system prompts). Minimum context size required before caching helps (usually 1K tokens). Caching works best for static context (documents) with varying queries. Prompt caching is increasingly available in modern APIs. Caching can dramatically reduce costs for some applications.

Last verified: 2026-04-08

Compare models

See how different LLMs compare on benchmarks, pricing, and speed.

Browse all models →