Skip to content

Session Pruning

Session pruning trims old tool results from the context before each LLM call. It reduces context bloat from accumulated tool outputs (exec results, file reads, search results) without touching your conversation messages.

Long sessions accumulate tool output that inflates the context window. This increases cost and can force compaction sooner than necessary.

Pruning is especially valuable for Anthropic prompt caching. After the cache TTL expires, the next request re-caches the full prompt. Pruning reduces the cache-write size, directly lowering cost.

  1. Wait for the cache TTL to expire (default 5 minutes).
  2. Find old tool results (user and assistant messages are never touched).
  3. Soft-trim oversized results — keep the head and tail, insert ....
  4. Hard-clear the rest — replace with a placeholder.
  5. Reset the TTL so follow-up requests reuse the fresh cache.

OpenClaw auto-enables pruning for Anthropic profiles:

Profile typePruning enabledHeartbeat
OAuth or setup-tokenYes1 hour
API keyYes30 min

If you set explicit values, OpenClaw does not override them.

Pruning is off by default for non-Anthropic providers. To enable:

{
agents: {
defaults: {
contextPruning: { mode: "cache-ttl", ttl: "5m" },
},
},
}

To disable: set mode: "off".

PruningCompaction
WhatTrims tool resultsSummarizes conversation
Saved?No (per-request)Yes (in transcript)
ScopeTool results onlyEntire conversation

They complement each other — pruning keeps tool output lean between compaction cycles.