prompt-caching

Installation
Summary

Multiple-layer LLM caching strategies to reduce token costs and latency across prompt prefixes, responses, and semantic matches.

  • Supports three caching approaches: Anthropic's native prompt caching for repeated prefixes, response caching for identical or similar queries, and Cache Augmented Generation (CAG) for pre-cached documents
  • Includes cache invalidation patterns and guidance on structuring prompts for optimal caching performance
  • Highlights critical anti-patterns: caching with high temperature, missing invalidation logic, and over-caching low-value data
  • Addresses sharp edges like cache miss latency spikes and prompt prefix changes that break caching effectiveness
SKILL.md

Prompt Caching

Caching strategies for LLM prompts including Anthropic prompt caching, response caching, and CAG (Cache Augmented Generation)

Capabilities

  • prompt-cache
  • response-cache
  • kv-cache
  • cag-patterns
  • cache-invalidation

Prerequisites

  • Knowledge: Caching fundamentals, LLM API usage, Hash functions
  • Skills_recommended: context-window-management

Scope

Related skills
Installs
533
GitHub Stars
37.3K
First Seen
Jan 19, 2026