Connect Clawsmith to your coding agent. Ship products like crazy.Unlimited usage during betaGet API Key →
← Back to dashboard
clawsmith.com/signal/openclaw-baseline-40k-tokens-per-message-workspace-injection
IssueUnderservedCoreLive

OpenClaw Baseline Token Bloat: 40K Tokens Per Message From Workspace File Injection

Every OpenClaw message incurs ~40K tokens of context loading regardless of reply length. Workspace files are re-injected into the system prompt on every API call, causing 93.5% token waste in multi-message conversations. Users report bills from $3,600/month (1.8M tokens). One system was observed sending 166K input tokens per call. No lazy-loading or first-message-only option exists.

Product Idea from this Signal

A plugin that prunes OpenClaw's 40K-token workspace injection down to only the files relevant to each message, cutting baseline API costs by 60-80%

217

OpenClaw injects the entire workspace context (SOUL.md, agent configs, skill descriptions, tool declarations) into every single message, consuming 40,000+ tokens before the user even types a word. This fixed overhead makes smaller models unusable (they hit context limits immediately) and multiplies API costs 3-5x for conversational sessions. This plugin intercepts the context assembly step and applies relevance filtering, only including workspace files and tool declarations that match the current message's intent, dropping the baseline from 40K to under 10K tokens.

PLUGINOPEN-SOURCECOST-OPTIMIZATIONCONTEXT-MANAGEMENT
CompetitiveView Opportunity →

Score Breakdown

Issues
217

Gap Assessment

UnderservedExisting solutions leave gaps

Community has proposed lazy-loading config but no official fix; third-party token optimizers like Lossless-Claw exist

Frequently Asked Questions