01
Prompt caching wired by default
System prompts and long context cached on every call. Cache hit rates above 90 percent on production workloads. Your token bill falls 60 to 80 percent versus a naive integration.
Cost per AI call drops from cents to fractions of a cent.