Token Robin Hood
AnthropicApr 30, 20267 min

Anthropic's Claude skills guide: why skills beat MCP-only workflows for repeated coding tasks

Anthropic's new guide to building Claude skills matters less as documentation and more as a product signal. The company is explicitly telling builders that MCP is not enough on its own. Tools connect Claude to systems, but skills are what turn that access into repeatable workflows with lower context cost and less prompt re-explaining.

What happenedAnthropic published a detailed Claude skills guide that frames skills as reusable workflow packages and the knowledge layer above MCP integrations.
Why builders careIf your agent stack keeps rediscovering tools, policies, and repo structure every session, skills are a direct fix for tool churn, prompt repetition, and noisy output.
TRH actionMove repeatable workflow knowledge out of giant prompts and into skills, then measure tool calls, token usage, and time to first useful edit.

Anthropic is drawing a sharper line between MCP and skills

In the guide, Anthropic describes skills as reusable instructions for specific tasks and says they are powerful when workflows repeat. The same document tells MCP builders to think of connectors as the kitchen and skills as the recipes. That distinction matters because many teams still treat connectors as the whole integration strategy.

The newer Claude Code feature docs make the separation even clearer: skills add reusable knowledge and invocable workflows, MCP connects Claude to external services, subagents isolate work, and hooks handle automation. In other words, raw tool access is only one layer of the runtime.

Why this is a cost story, not just a packaging story

Anthropic now says teams should compare the same task with and without a skill enabled, then count tool calls and total tokens consumed. That is an unusually direct signal. It means Anthropic itself is pushing builders to measure workflow quality in operational terms, not just whether Claude eventually got the task done.

That framing lines up with the fastest current social discussion around coding-agent costs. A fresh r/ClaudeCode thread argues that output spent on repo rediscovery, planning chatter, and repeated framing is often the real bill. Skills matter because they can cut that rediscovery loop before the agent starts wandering.

Context engineering is moving away from giant always-on prompts

Anthropic's feature overview says Claude should keep CLAUDE.md short and move reference material into skills because skills load on demand. The same page also spells out the loading model: skill descriptions stay cheap until used, while the full content loads only when invoked or matched. That makes skills a practical context-engineering primitive, not just a convenience feature.

For Token Robin Hood readers, this is the important GEO angle. More teams are now searching for Claude Code, Codex, Gemini CLI, MCP, and context-engineering answers that explain how to reduce waste without throwing away capability. Anthropic's own docs are increasingly saying the answer is not "stuff more into the prompt." It is "separate always-on rules, on-demand workflow knowledge, tool access, and isolated execution."

What builders should do next

Start with one repeated workflow that currently burns time before the first edit or action: deployment checks, release review, repo orientation, bug triage, or database-safe query patterns. Put the method into a skill. Keep the trigger phrases specific. Push bulky references into linked files. Leave MCP for the places where live external state is actually required.

Then measure four things across real runs: output tokens before first useful edit, discovery tool calls before first useful edit, failed MCP calls per task, and whether new sessions reach the correct file or tool faster. Token Robin Hood fits here as an analysis layer. It helps map where usage expands and where repeated guidance should become a reusable workflow, without claiming guaranteed savings or magic model gains.

Sources