Token Robin Hood
Claude CodeApr 30, 20267 min

Claude Code update: MCP hooks, forked subagents, and a more honest 1M context meter

Anthropic's Week 17 Claude Code release is not just another polish pass. The meaningful builder changes sit lower in the stack: hooks can now call MCP tools directly, forked subagents can carry full context into external builds, and Opus 4.7 sessions now compute against the model's native 1M context window instead of overstating pressure.

What happenedAnthropic published a fresh Claude Code update for April 20-24, 2026 with direct MCP hooks, forked-subagent support on external builds, faster resume behavior, and cleaner context accounting.
Why builders careThese changes cut orchestration friction exactly where coding-agent teams lose time and tokens: runtime guardrails, context handoff, and noisy session measurement.
TRH actionRoute deterministic checks through hooks, use subagents only for bounded work, and start measuring output before first edit instead of only prompt size.

Hooks are becoming a real control plane

The biggest signal is that hooks can now call MCP tools directly through type: "mcp_tool". That removes one layer of shell glue when teams want enforcement, logging, approvals, or policy checks around existing connected tools. In practice, Claude Code is moving from "agent with tools" toward "agent runtime with programmable guardrails."

That lines up with Anthropic's own advanced-patterns material from March 24, which framed hooks as the place for deterministic lifecycle automation and MCP as the layer for reasoning over external systems. When those two layers connect more directly, builders can spend less output budget narrating tool choreography and more budget on the actual task.

Forked subagents are useful only if you keep roles tight

Anthropic also exposed forked subagents on external builds behind CLAUDE_CODE_FORK_SUBAGENT=1, which lets a fork inherit the full conversation instead of starting from zero. That is powerful, but it also raises the risk that teams use subagents as a blanket answer to every hard task.

The official subagent docs and webinar material both push the same constraint: subagents work best when they have a clear role, narrow tool access, and a small return artifact. That matches the latest social signal too. A fresh r/ClaudeCode thread argues that many coding-agent bills are dominated by output spent on repo rediscovery, planning chatter, and repeated framing before the first useful edit. Forking more agents without tighter task boundaries can make that worse, not better.

The 1M context fix matters because fake pressure creates fake optimization

Anthropic says Opus 4.7 sessions now compute against the model's native 1M context window, fixing inflated /context percentages and premature autocompaction. That matters because bad meters create bad operator behavior. If a team thinks the window is fuller than it really is, they start summarizing too early, compacting too aggressively, or splitting tasks in ways that add more coordination chatter than they remove.

Search Console already shows that Claude Code 2026 attracts impressions for Token Robin Hood readers. This update sharpens the practical angle: the context window is not just model trivia. It is a runtime budget surface that changes how teams decide to branch, summarize, delegate, and review.

What Token Robin Hood readers should do now

Move deterministic behavior into hooks first. Use MCP where the agent genuinely needs external state, not where a local script is enough. Keep subagents reserved for bounded investigation, review, or parallel slices with explicit success criteria. Then measure three things across real sessions: output tokens before first edit, discovery tool calls before first edit, and whether the agent reached the correct file earlier.

That is the useful line between automation and waste. Token Robin Hood can help analyze where token usage expands, where context gets replayed, and where orchestration turns into invisible spend. The claim is not guaranteed savings. The claim is better visibility into the exact runtime habits that decide whether a stronger agent stack becomes faster or just louder.

Sources