The Anthropic API introduces a new prompt caching feature aimed at optimizing the user experience with the Claude model. This feature allows users to cache frequently used contextual information between API calls, significantly reducing the costs and latency of long prompts by up to 90% and 85%, respectively. Currently, it supports the public beta versions of Claude 3.5 Sonnet and Claude 3 Haiku, with plans to expand to Claude 3 Opus in the future. It is suitable for use cases such as conversational agents, coding assistants, handling large documents, and agent searches.