prompt-caching
Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augm...
Best use case
prompt-caching is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augm...
Teams using prompt-caching should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/prompt-caching/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How prompt-caching Compares
| Feature / Agent | prompt-caching | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation) Use when: prompt caching, cache prompt, response cache, cag, cache augm...
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Prompt Caching You're a caching specialist who has reduced LLM costs by 90% through strategic caching. You've implemented systems that cache at multiple levels: prompt prefixes, full responses, and semantic similarity matches. You understand that LLM caching is different from traditional caching—prompts have prefixes that can be cached, responses vary with temperature, and semantic similarity often matters more than exact match. Your core principles: 1. Cache at the right level—prefix, response, or both 2. K ## Capabilities - prompt-cache - response-cache - kv-cache - cag-patterns - cache-invalidation ## Patterns ### Anthropic Prompt Caching Use Claude's native prompt caching for repeated prefixes ### Response Caching Cache full LLM responses for identical or similar queries ### Cache Augmented Generation (CAG) Pre-cache documents in prompt instead of RAG retrieval ## Anti-Patterns ### ❌ Caching with High Temperature ### ❌ No Cache Invalidation ### ❌ Caching Everything ## ⚠️ Sharp Edges | Issue | Severity | Solution | |-------|----------|----------| | Cache miss causes latency spike with additional overhead | high | // Optimize for cache misses, not just hits | | Cached responses become incorrect over time | high | // Implement proper cache invalidation | | Prompt caching doesn't work due to prefix changes | medium | // Structure prompts for optimal caching | ## Related Skills Works well with: `context-window-management`, `rag-implementation`, `conversation-memory` ## When to Use This skill is applicable to execute the workflow or actions described in the overview.
Related Skills
tldr-prompt
Create tldr summaries for GitHub Copilot files (prompts, agents, instructions, collections), MCP servers, or documentation from URLs and queries.
customaize-agent:test-prompt
Use when creating or editing any prompt (commands, hooks, skills, subagent instructions) to verify it produces desired behavior - applies RED-GREEN-REFACTOR cycle to prompt engineering using subagents for isolated testing
prompt-library
Curated collection of high-quality prompts for various use cases. Includes role-based prompts, task-specific templates, and prompt refinement techniques. Use when user needs prompt templates, role-...
customaize-agent:prompt-engineering
Use this skill when you writing commands, hooks, skills for Agent, or prompts for sub agents or any other LLM interaction, including optimizing prompts, improving LLM outputs, or designing production prompt templates.
prompt-engineering-patterns
Master advanced prompt engineering techniques to maximize LLM performance, reliability, and controllability in production. Use when optimizing prompts, improving LLM outputs, or designing productio...
prompt-engineer
Transforms user prompts into optimized prompts using frameworks (RTF, RISEN, Chain of Thought, RODES, Chain of Density, RACE, RISE, STAR, SOAP, CLEAR, GROW)
prompt-builder
Guide users through creating high-quality GitHub Copilot prompts with proper structure, tools, and best practices.
llm-application-dev-prompt-optimize
You are an expert prompt engineer specializing in crafting effective prompts for LLMs through advanced techniques including constitutional AI, chain-of-thought reasoning, and model-specific optimizati
finalize-agent-prompt
Finalize prompt file using the role of an AI agent to polish the prompt for the end user.
enhance-prompts
Use when improving general prompts for structure, examples, and constraints.
enhance-agent-prompts
Use when improving agent prompts, frontmatter, and tool restrictions.
boost-prompt
Interactive prompt refinement workflow: interrogates scope, deliverables, constraints; copies final markdown to clipboard; never writes code. Requires the Joyride extension.