Add prompt caching for ~80% input token savings
Implements Anthropic's prompt caching to reduce API costs. The system
now uses three cache breakpoints:
- Tools (cache_control on last tool definition)
- System prompt split into base prompt + context blocks
- Messages (cache_control on last message each turn)
Also switches default model to Sonnet 4.5 since it has a lower cache
minimum (1,024 tokens vs Haiku's 4,096).
Adds `--debug` flag to `storied play` that shows token usage after each
response, including cache hits/writes. The `/context` command also now
shows cache stats in session totals.
See https://platform.claude.com/docs/en/build-with-claude/prompt-caching
馃 Generated with [Claude Code](https://claude.com/claude-code)
Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>