Package Exports
- @tarquinen/opencode-dcp
- @tarquinen/opencode-dcp/dist/index.js
This package does not declare an exports field, so the exports above have been automatically detected and optimized by JSPM instead. If any package subpath is missing, it is recommended to post an issue to the original package (@tarquinen/opencode-dcp) to support the "exports" field. If that is not possible, create a JSPM override to customize the exports field for this package.
Readme
Dynamic Context Pruning Plugin
Automatically reduces token usage in OpenCode by removing obsolete tool outputs from conversation history.

Installation
Add to your OpenCode config:
// opencode.jsonc
{
"plugin": ["@tarquinen/opencode-dcp@0.3.20"]
}When a new version is available, DCP will show a toast notification. Update by changing the version number in your config.
Restart OpenCode. The plugin will automatically start optimizing your sessions.
Pruning Strategies
DCP implements two complementary strategies:
Deduplication — Fast, zero-cost pruning that identifies repeated tool calls (e.g., reading the same file multiple times) and keeps only the most recent output. Runs instantly with no LLM calls.
AI Analysis — Uses a language model to semantically analyze conversation context and identify tool outputs that are no longer relevant to the current task. More thorough but incurs LLM cost.
How It Works
DCP is non-destructive—pruning state is kept in memory only. When requests go to your LLM, DCP replaces pruned outputs with a placeholder; original session data stays intact.
Configuration
DCP uses its own config file (~/.config/opencode/dcp.jsonc or .opencode/dcp.jsonc), created automatically on first run.
Options
| Option | Default | Description |
|---|---|---|
enabled |
true |
Enable/disable the plugin |
debug |
false |
Log to ~/.config/opencode/logs/dcp/ |
model |
(session) | Model for analysis (e.g., "anthropic/claude-haiku-4-5") |
showModelErrorToasts |
true |
Show notifications on model fallback |
strictModelSelection |
false |
Only run AI analysis with session or configured model (disables fallback models) |
pruning_summary |
"detailed" |
"off", "minimal", or "detailed" |
protectedTools |
["task", "todowrite", "todoread", "context_pruning"] |
Tools that are never pruned |
strategies.onIdle |
["deduplication", "ai-analysis"] |
Strategies for automatic pruning |
strategies.onTool |
["deduplication", "ai-analysis"] |
Strategies when AI calls context_pruning |
Strategies: "deduplication" (fast, zero LLM cost) and "ai-analysis" (maximum savings). Empty array disables that trigger.
{
"enabled": true,
"strategies": {
"onIdle": ["deduplication", "ai-analysis"],
"onTool": ["deduplication", "ai-analysis"]
},
"protectedTools": ["task", "todowrite", "todoread", "context_pruning"]
}Config Precedence
Settings are merged in order: Defaults → Global (~/.config/opencode/dcp.jsonc) → Project (.opencode/dcp.jsonc). Each level overrides the previous, so project settings take priority over global, which takes priority over defaults.
Restart OpenCode after making config changes.
License
MIT