You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Automatically reduces token usage in OpenCode by removing obsolete tools from conversation history.
6
+
Automatically reduces token usage in OpenCode by removing obsolete content from conversation history.
7
7
8
8

9
9
@@ -50,7 +50,7 @@ LLM providers like Anthropic and OpenAI cache prompts based on exact prefix matc
50
50
51
51
**Trade-off:** You lose some cache read benefits but gain larger token savings from reduced context size and performance improvements through reduced context poisoning. In most cases, the token savings outweigh the cache miss cost—especially in long sessions where context bloat becomes significant.
52
52
53
-
> **Note:** In testing, cache hit rates were approximately 65% with DCP enabled vs 85% without.
53
+
> **Note:** In testing, cache hit rates were approximately 80% with DCP enabled vs 85% without for most providers.
54
54
55
55
**Best use case:** Providers that count usage in requests, such as Github Copilot and Google Antigravity, have no negative price impact.
0 commit comments