Package Exports
- @ramtinj95/opencode-tokenscope
- @ramtinj95/opencode-tokenscope/dist/tokenscope.js
This package does not declare an exports field, so the exports above have been automatically detected and optimized by JSPM instead. If any package subpath is missing, it is recommended to post an issue to the original package (@ramtinj95/opencode-tokenscope) to support the "exports" field. If that is not possible, create a JSPM override to customize the exports field for this package.
Readme
OpenCode-Tokenscope, Token Analyzer Plugin
Comprehensive token usage analysis and cost tracking for OpenCode AI sessions
Track and optimize your token usage across system prompts, user messages, tool outputs, and more. Get detailed breakdowns, accurate cost estimates, and visual insights for your AI development workflow.
Installation
curl -sSL https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/install.sh | bashThen restart OpenCode and run /tokenscope
Updating
Option 1: Local script (if you have the plugin installed)
bash ~/.config/opencode/plugin/install.sh --updateOption 2: Remote script (always works)
curl -sSL https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/install.sh | bash -s -- --updateThe --update flag skips dependency installation for faster updates.
Usage
Simply type in OpenCode:
/tokenscopeThe plugin will:
- Analyze the current session
- Count tokens across all categories
- Analyze all subagent (Task tool) child sessions recursively
- Calculate costs based on API telemetry
- Save detailed report to
token-usage-output.txt
Options
- sessionID: Analyze a specific session instead of the current one
- limitMessages: Limit entries shown per category (1-10, default: 3)
- includeSubagents: Include subagent child session costs (default: true)
Reading the Full Report
cat token-usage-output.txtFeatures
Comprehensive Token Analysis
- 5 Category Breakdown: System prompts, user messages, assistant responses, tool outputs, and reasoning traces
- Visual Charts: Easy-to-read ASCII bar charts with percentages and token counts
- Smart Inference: Automatically infers system prompts from API telemetry (since they're not exposed in session messages)
Context Breakdown Analysis (New in v1.4.0)
- System Prompt Components: See token distribution across base prompt, tool definitions, environment context, project tree, and custom instructions
- Automatic Estimation: Estimates breakdown from
cache_writetokens when system prompt content isn't directly available - Tool Count: Shows how many tools are loaded and their combined token cost
Tool Definition Cost Estimates (New in v1.4.0)
- Per-Tool Estimates: Lists all enabled tools with estimated schema token costs
- Argument Analysis: Infers argument count and complexity from actual tool calls in the session
- Complexity Detection: Distinguishes between simple arguments and complex ones (arrays/objects)
Cache Efficiency Metrics (New in v1.4.0)
- Cache Hit Rate: Visual display of cache read vs fresh input token distribution
- Cost Savings: Calculates actual savings from prompt caching
- Effective Rate: Shows what you're actually paying per token vs standard rates
Accurate Cost Tracking
- 41+ Models Supported: Comprehensive pricing database for Claude, GPT, DeepSeek, Llama, Mistral, and more
- Cache-Aware Pricing: Properly handles cache read/write tokens with discounted rates
- Session-Wide Billing: Aggregates costs across all API calls in your session
Subagent Cost Tracking
- Child Session Analysis: Recursively analyzes all subagent sessions spawned by the Task tool
- Aggregated Totals: Shows combined tokens, costs, and API calls across main session and all subagents
- Per-Agent Breakdown: Lists each subagent with its type, token usage, cost, and API call count
- Optional Toggle: Enable/disable subagent analysis with the
includeSubagentsparameter
Advanced Features
- Tool Usage Stats: Track which tools consume the most tokens and how many times each is called
- API Call Tracking: See total API calls for main session and subagents
- Top Contributors: Identify the biggest token consumers
- Model Normalization: Handles
provider/modelformat automatically - Multi-Tokenizer Support: Uses official tokenizers (tiktoken for OpenAI, transformers for others)
- Configurable Sections: Enable/disable analysis features via
tokenscope-config.json
Example Output
═══════════════════════════════════════════════════════════════════════════
Token Analysis: Session ses_50c712089ffeshuuuJPmOoXCPX
Model: claude-opus-4-5
═══════════════════════════════════════════════════════════════════════════
TOKEN BREAKDOWN BY CATEGORY
─────────────────────────────────────────────────────────────────────────
Estimated using tokenizer analysis of message content:
Input Categories:
SYSTEM ██████████████░░░░░░░░░░░░░░░░ 45.8% (22,367)
USER ░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░ 0.8% (375)
TOOLS ████████████████░░░░░░░░░░░░░░ 53.5% (26,146)
Subtotal: 48,888 estimated input tokens
Output Categories:
ASSISTANT ██████████████████████████████ 100.0% (1,806)
Subtotal: 1,806 estimated output tokens
Local Total: 50,694 tokens (estimated)
TOOL USAGE BREAKDOWN
─────────────────────────────────────────────────────────────────────────
bash ██████████░░░░░░░░░░░░░░░░░░░░ 34.0% (8,886) 4x
read ██████████░░░░░░░░░░░░░░░░░░░░ 33.1% (8,643) 3x
task ████████░░░░░░░░░░░░░░░░░░░░░░ 27.7% (7,245) 4x
webfetch █░░░░░░░░░░░░░░░░░░░░░░░░░░░░░ 4.9% (1,286) 1x
tokenscope ░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░ 0.3% (75) 2x
batch ░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░ 0.0% (11) 1x
TOP CONTRIBUTORS
─────────────────────────────────────────────────────────────────────────
• System (inferred from API) 22,367 tokens (44.1%)
• bash 8,886 tokens (17.5%)
• read 8,643 tokens (17.0%)
• task 7,245 tokens (14.3%)
• webfetch 1,286 tokens (2.5%)
═══════════════════════════════════════════════════════════════════════════
MOST RECENT API CALL
─────────────────────────────────────────────────────────────────────────
Raw telemetry from last API response:
Input (fresh): 2 tokens
Cache read: 48,886 tokens
Cache write: 54 tokens
Output: 391 tokens
───────────────────────────────────
Total: 49,333 tokens
═══════════════════════════════════════════════════════════════════════════
SESSION TOTALS (All 15 API calls)
─────────────────────────────────────────────────────────────────────────
Total tokens processed across the entire session (for cost calculation):
Input tokens: 10 (fresh tokens across all calls)
Cache read: 320,479 (cached tokens across all calls)
Cache write: 51,866 (tokens written to cache)
Output tokens: 3,331 (all model responses)
───────────────────────────────────
Session Total: 375,686 tokens (for billing)
═══════════════════════════════════════════════════════════════════════════
ESTIMATED SESSION COST (API Key Pricing)
─────────────────────────────────────────────────────────────────────────
You appear to be on a subscription plan (API cost is $0).
Here's what this session would cost with direct API access:
Input tokens: 10 × $5.00/M = $0.0001
Output tokens: 3,331 × $25.00/M = $0.0833
Cache read: 320,479 × $0.50/M = $0.1602
Cache write: 51,866 × $6.25/M = $0.3242
─────────────────────────────────────────────────────────────────────────
ESTIMATED TOTAL: $0.5677
Note: This estimate uses standard API pricing from models.json.
Actual API costs may vary based on provider and context size.
═══════════════════════════════════════════════════════════════════════════
CONTEXT BREAKDOWN (Estimated from cache_write tokens)
─────────────────────────────────────────────────────────────────────────
Base System Prompt ████████████░░░░░░░░░░░░░░░░░░ ~42,816 tokens
Tool Definitions (14)██████░░░░░░░░░░░░░░░░░░░░░░░░ ~4,900 tokens
Environment Context █░░░░░░░░░░░░░░░░░░░░░░░░░░░░░ ~150 tokens
Project Tree ████░░░░░░░░░░░░░░░░░░░░░░░░░░ ~4,000 tokens
───────────────────────────────────────────────────────────────────────
Total Cached Context: ~51,866 tokens
Note: Breakdown estimated from first cache_write. Actual distribution may vary.
═══════════════════════════════════════════════════════════════════════════
TOOL DEFINITION COSTS (Estimated from argument analysis)
─────────────────────────────────────────────────────────────────────────
Tool Est. Tokens Args Complexity
───────────────────────────────────────────────────────────────────────
task ~480 3 complex (arrays/objects)
batch ~410 1 complex (arrays/objects)
edit ~370 4 simple
read ~340 3 simple
bash ~340 3 simple
───────────────────────────────────────────────────────────────────────
Total: ~4,520 tokens (14 enabled tools)
Note: Estimates inferred from tool call arguments in this session.
Actual schema tokens may vary +/-20%.
═══════════════════════════════════════════════════════════════════════════
CACHE EFFICIENCY
─────────────────────────────────────────────────────────────────────────
Token Distribution:
Cache Read: 320,479 tokens ████████████████████████████░░ 86.2%
Fresh Input: 51,320 tokens ████░░░░░░░░░░░░░░░░░░░░░░░░░░ 13.8%
───────────────────────────────────────────────────────────────────────
Cache Hit Rate: 86.2%
Cost Analysis (claude-opus-4-5 @ $5.00/M input, $0.50/M cache read):
Without caching: $1.8590 (371,799 tokens x $5.00/M)
With caching: $0.4169 (fresh x $5.00/M + cached x $0.50/M)
───────────────────────────────────────────────────────────────────────
Cost Savings: $1.4421 (77.6% reduction)
Effective Rate: $1.12/M tokens (vs. $5.00/M standard)
═══════════════════════════════════════════════════════════════════════════
SUBAGENT COSTS (4 child sessions, 23 API calls)
─────────────────────────────────────────────────────────────────────────
docs $0.3190 (194,701 tokens, 8 calls)
general $0.2957 (104,794 tokens, 4 calls)
docs $0.2736 (69,411 tokens, 4 calls)
general $0.5006 (197,568 tokens, 7 calls)
─────────────────────────────────────────────────────────────────────────
Subagent Total: $1.3888 (566,474 tokens, 23 calls)
═══════════════════════════════════════════════════════════════════════════
SUMMARY
─────────────────────────────────────────────────────────────────────────
Cost Tokens API Calls
Main session: $ 0.5677 375,686 15
Subagents: $ 1.3888 566,474 23
─────────────────────────────────────────────────────────────────────────
TOTAL: $ 1.9565 942,160 38
═══════════════════════════════════════════════════════════════════════════
Supported Models
41+ models with accurate pricing:
Claude Models
- Claude Opus 4.5, 4.1, 4
- Claude Sonnet 4, 4-5, 3.7, 3.5, 3
- Claude Haiku 4-5, 3.5, 3
OpenAI Models
- GPT-4, GPT-4 Turbo, GPT-4o, GPT-4o Mini
- GPT-3.5 Turbo
- GPT-5 and all its variations
Other Models
- DeepSeek (R1, V2, V3)
- Llama (3.1, 3.2, 3.3)
- Mistral (Large, Small)
- Qwen, Kimi, GLM, Grok
- And more...
Free/Open models are marked with zero pricing.
Customization
Add New Model Pricing
Edit ~/.config/opencode/plugin/models.json:
{
"your-model-name": {
"input": 1.50,
"output": 5.00,
"cacheWrite": 0.50,
"cacheRead": 0.10
}
}Save the file and restart OpenCode. The plugin will automatically use the new pricing.
Update Existing Model Pricing
Simply edit the values in models.json and restart OpenCode. No code changes needed!
Configure Analysis Features
Edit ~/.config/opencode/plugin/tokenscope-config.json to enable/disable sections:
{
"enableContextBreakdown": true,
"enableToolSchemaEstimation": true,
"enableCacheEfficiency": true,
"enableSubagentAnalysis": true
}Set any option to false to hide that section from the output. All features are enabled by default.
How It Works
System Prompt Inference
OpenCode doesn't expose system prompts in the session messages API. The plugin intelligently infers them using:
System Tokens = (API Input + Cache Read) - (User Tokens + Tool Tokens)This works because the API input includes everything sent to the model.
Dual Tracking
- Current Context: Uses the most recent API call with non-zero tokens (matches TUI)
- Session Total: Aggregates all API calls for accurate billing
Subagent Analysis
The plugin uses OpenCode's session API to:
- Fetch all child sessions spawned by the Task tool
- Recursively analyze nested subagents (subagents can spawn their own subagents)
- Aggregate tokens, costs, and API call counts
- Calculate estimated costs using the same pricing as the main session
Model Name Normalization
Automatically handles provider/model format (e.g., qwen/qwen3-coder → qwen3-coder)
Understanding the Numbers
Current Context vs Session Total
Current Context: What's in your context window right now
- Based on most recent API call
- Used to understand current memory usage
Session Total: All tokens processed in this session
- Sum of all API calls in the main session
- What you're billed for (main session only)
- Used for cost calculation
Subagent Totals
When using the Task tool, OpenCode spawns subagent sessions. These are tracked separately:
- Subagent Tokens: Combined tokens from all child sessions
- Subagent API Calls: Total API calls made by all subagents
- Grand Total: Main session + all subagents combined
Cache Tokens
- Cache Read: Tokens retrieved from cache (discounted rate ~90% off)
- Cache Write: Tokens written to cache (slight premium ~25% more)
- Note: Cache write is a billing charge, not additional context tokens
Troubleshooting
"Dependencies missing" Error
Run the installer:
curl -sSL https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/install.sh | bashCommand Not Appearing
- Verify
tokenscope.mdexists:ls ~/.config/opencode/command/tokenscope.md - Restart OpenCode completely
- Check OpenCode logs for plugin errors
Wrong Token Counts
The plugin uses API telemetry (ground truth). If counts seem off:
- Expected ~2K difference from TUI: Plugin analyzes before its own response is added
- Model detection: Check that the model name is recognized in the output
- Tokenizer not installed: Re-run the installer
New Model Not Showing Correct Pricing
- Check if model exists in
models.json - Try exact match or prefix match (e.g.,
claude-sonnet-4matchesclaude-sonnet-4-20250514) - Add entry to
models.jsonif missing - Restart OpenCode after editing
models.json
Plugin Fails to Load
- Validate JSON syntax:
cd ~/.config/opencode/plugin node -e "JSON.parse(require('fs').readFileSync('models.json', 'utf8'))"
- Check for trailing commas or syntax errors
- Plugin falls back to default pricing if file is invalid
Architecture
File Structure
plugin/
├── tokenscope.ts # Main entry point - Plugin export
├── tokenscope-lib/
│ ├── types.ts # All interfaces and type definitions
│ ├── config.ts # Constants, model maps, pricing loader
│ ├── tokenizer.ts # TokenizerManager class
│ ├── analyzer.ts # ModelResolver, ContentCollector, TokenAnalysisEngine
│ ├── cost.ts # CostCalculator class
│ ├── subagent.ts # SubagentAnalyzer class
│ ├── formatter.ts # OutputFormatter class
│ └── context.ts # ContextAnalyzer class (context breakdown, tool estimates, cache efficiency)
├── models.json # Pricing data for 41+ models
├── tokenscope-config.json # Feature toggles configuration
├── package.json # Plugin metadata
└── install.sh # Installation scriptCore Components
- TokenizerManager (
tokenscope-lib/tokenizer.ts): Loads and caches tokenizers (tiktoken, transformers) - ModelResolver (
tokenscope-lib/analyzer.ts): Detects model and selects appropriate tokenizer - ContentCollector (
tokenscope-lib/analyzer.ts): Extracts content from session messages, including tool call counts - TokenAnalysisEngine (
tokenscope-lib/analyzer.ts): Counts tokens and applies API telemetry adjustments - CostCalculator (
tokenscope-lib/cost.ts): Calculates costs from pricing database with cache-aware pricing - SubagentAnalyzer (
tokenscope-lib/subagent.ts): Recursively fetches and analyzes child sessions from Task tool calls - ContextAnalyzer (
tokenscope-lib/context.ts): Analyzes context breakdown, tool schema estimates, and cache efficiency - OutputFormatter (
tokenscope-lib/formatter.ts): Generates visual reports with charts and summaries
Privacy & Security
- All processing is local: No session data sent to external services
- Tokenizers from official sources:
- OpenAI tokenizers: npm registry
- Transformers: Hugging Face Hub
- Open source: Audit the code yourself
Performance
- Fast: Tokenizers cached after first load
- Parallel: Categories processed concurrently
- Efficient: Only analyzes on demand
- First-run download: Transformers models download on demand (5-50MB per model)
- Subsequent runs: Instant (uses cache)
Manual Installation
Click to expand manual installation steps
Requirements
- OpenCode installed (
~/.config/opencodedirectory exists) - npm (for tokenizer dependencies)
- ~50MB disk space (for tokenizer models)
Installation Steps
Navigate to OpenCode config:
cd ~/.config/opencodeDownload plugin files:
mkdir -p plugin/tokenscope-lib cd plugin curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/models.json curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/package.json curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-config.json cd tokenscope-lib curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/types.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/config.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/tokenizer.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/analyzer.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/cost.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/subagent.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/formatter.ts curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/plugin/tokenscope-lib/context.ts
Download command file:
cd ../../command curl -O https://raw.githubusercontent.com/ramtinJ95/opencode-tokenscope/main/command/tokenscope.md
Install dependencies:
cd ../plugin npm install js-tiktoken@1.0.15 @huggingface/transformers@3.1.2
Restart OpenCode
Test: Run
/tokenscopein any session
Contributing
Contributions welcome! Ideas for enhancement:
- Historical trend analysis
- Export to CSV/JSON/PDF
- Optimization suggestions
- Custom categorization rules
- Real-time monitoring with alerts
- Compare sessions
- Token burn rate calculation
Support
- Issues: GitHub Issues
- Discussions: GitHub Discussions