Enables developers to cache frequently used context between API calls, providing Claude with more background knowledge and example outputs.
Now available on the Anthropic API
Conversational agents: Reduce cost and latency for extended conversations
Coding assistants: Improve autocomplete and codebase Q&A
Large document processing: Incorporate complete long-form material including images
Talk to books and papers: Bring any knowledge base alive for Q&A
Action | Cost |
---|---|
Writing to cache | 25% more than base input token price |
Using cached content | Only 10% of base input token price |
Notion is adding prompt caching to Claude-powered features for Notion AI
"We're excited to use prompt caching to make Notion AI faster and cheaper, all while maintaining state-of-the-art quality."
— Simon Last, Co-founder at Notion
@Anthropic