LogoCua Documentation

Prompt Caching

How to use prompt caching in ComputerAgent and agent loops.

Prompt caching is a cost-saving feature offered by some LLM API providers that helps avoid reprocessing the same prompt, improving efficiency and reducing costs for repeated or long-running tasks.

Usage

The use_prompt_caching argument is available for ComputerAgent and agent loops:

agent = ComputerAgent(
    ...,
    use_prompt_caching=True,
)
  • Type: bool
  • Default: False
  • Purpose: Use prompt caching to avoid reprocessing the same prompt.

Anthropic CUAs

When using Anthropic-based CUAs (Claude models), setting use_prompt_caching=True will automatically add { "cache_control": "ephemeral" } to your messages. This enables prompt caching for the session and can speed up repeated runs with the same prompt.

Note: This argument is only required for Anthropic CUAs. For other providers, it is ignored.

OpenAI Provider

With the OpenAI provider, prompt caching is handled automatically for prompts of 1000+ tokens. You do not need to set use_prompt_caching—caching will occur for long prompts without any extra configuration.

Example

from agent import ComputerAgent
agent = ComputerAgent(
    model="anthropic/claude-3-5-sonnet-20240620",
    use_prompt_caching=True,
)

Implementation Details

  • For Anthropic: Adds { "cache_control": "ephemeral" } to messages when enabled.
  • For OpenAI: Caching is automatic for long prompts; the argument is ignored.

When to Use

  • Enable for Anthropic CUAs if you want to avoid reprocessing the same prompt in repeated or iterative tasks.
  • Not needed for OpenAI models unless you want explicit ephemeral cache control (not required for most users).

See Also