Template L · Concept · D3 Agent Operations

Skills.

7 min read·10 sections·Tier A

Skills are reusable, on-demand task workflows defined in markdown with YAML frontmatter. Unlike CLAUDE.md (always loaded), skills load only when Claude matches the description to the current task. Skills can restrict tool access and accept arguments.

On-demand automationDomain 3Match-by-description
Skills — hero illustration featuring Loop mascot in a warm gallery scene.
Domain D3Agent Operations · 20%
On this page
01 · Summary

TLDR

Skills are reusable, on-demand task workflows defined in markdown with YAML frontmatter. Unlike CLAUDE.md (always loaded), skills load only when Claude matches the description to the current task. Skills can restrict tool access and accept arguments.

5
Frontmatter fields
2
Scopes
on-demand
Load pattern
D3
Exam domain
fork
Context isolation
02 · Definition

What it is

A skill is a self-contained folder of instructions that Claude Code discovers and activates automatically when it recognizes the task. Instead of repeating "review this PR for security" or "format commits this way" across sessions, you write a SKILL.md file once, with a name, description, and instructions in YAML frontmatter, and Claude applies it on demand. The mental model: a skill is a task-specific, portable teaching file that lives in ~/.claude/skills/ (personal) or .claude/skills/ (project, version-controlled, shared).

What makes a skill skill-like (not just a CLAUDE.md or a slash command) is automatic matching and progressive disclosure. CLAUDE.md always loads, burning tokens on rules you don't need. Slash commands require explicit invocation every time. Skills are different: Claude reads each skill's description, compares it against your request, and activates only the relevant ones. No token overhead when idle; no manual triggering. The description is the selector.

Skills support progressive disclosure through multi-file organization. Keep the core SKILL.md under 500 lines, then link to supporting files (references/, scripts/, assets/) that Claude reads only when needed. Scripts execute without loading their contents into context, only the output consumes tokens. A skill can restrict tool access via allowed-tools, perfect for read-only exploration or security-sensitive work where you don't want Claude to accidentally modify files.

The skill frontmatter pattern is minimalist by design. Two required fields: name (lowercase, hyphens, max 64 chars) and description (max 1,024 chars; this is where matching magic lives). Two optional: allowed-tools and model. Below the frontmatter, you write instructions in plain Markdown. Adoption reflex: "I'm repeating this to Claude across sessions" → a skill is waiting to be written.

03 · Mechanics

How it works

The discovery flow has four stages. Matching: Claude reads your request and compares it semantically against the descriptions of all available skills (personal and project). If your request mentions "review PR" and a skill says "reviews pull requests for code quality," it's a match. Loading: Claude loads only the matched skill's SKILL.md into context, not all skills. Execution: you interact with the skill (Claude references its instructions). Unloading: when the task finishes, the skill context is discarded.

Skill matching is semantic and probabilistic. Claude uses natural language understanding, not string matching. If your request is "please check this code change" and a skill describes itself as "ensures code quality in pull requests," Claude can match them even though exact words differ. This is why descriptions matter so much: vague descriptions like "helps with code" don't match reliably; explicit ones like "validates Python security patterns" consistently trigger.

Tool access is restricted at activation time, not at request time. When you set allowed-tools: Read, Grep, Glob, Bash in the frontmatter, Claude can only use those tools while the skill is active. It cannot call Edit or Write, even if you ask. Perfect for a "codebase-onboarding" skill where new developers should explore but not modify. If you omit allowed-tools, the skill inherits your normal permission model.

Progressive disclosure saves context budget. A big skill with a 3,000-line reference document burns 2,000 tokens just sitting in context. Instead, keep SKILL.md lean, then structure: skill-dir/references/guide.md, skill-dir/scripts/validate.sh. In SKILL.md, write: "if the user asks about X, run scripts/validate.sh and read references/guide.md." Claude loads only what's needed. Scripts execute in the background; their output (not their source) gets injected.

Skills mechanics, painterly diagram featuring Loop mascot.
04 · In production

Where you'll see it

Domain-specific research workflow

Healthcare team has /healthcare-research skill that searches PubMed + extracts claims with sources + tags evidence tier (RCT vs observational). Description matches 'research medical' or 'clinical evidence' queries. Tools restricted to Read/WebSearch/WebFetch, no Edit, no Bash. Output is JSON.

Headless CI code review

/ci-review skill runs in claude -p headless mode from GitHub Actions. Input is a PR diff; output is structured JSON of findings. context: fork keeps the verbose review out of the main session. Same skill works locally and in CI without changes.

Vault search with depth control

/search-vault skill takes a query + depth argument. At depth=1 it returns frontmatter snippets; at depth=3 it returns full sections. Skill matches 'find in vault' / 'search notes' descriptions. Reduces context bloat by serving the right depth instead of dumping full files.

05 · Implementation

Code examples

.claude/skills/healthcare-research.mdyaml
---
name: healthcare-research
description: |
  Use when the user asks for medical literature, clinical evidence,
  drug efficacy data, or similar healthcare-research queries.
  Returns structured JSON: {claims: [{claim, sources: [...], tier}]}.
context: fork
allowed-tools: [Read, WebSearch, WebFetch]
argument-hint: "Search query (e.g., 'CDK4/6 inhibitors HR+ breast cancer')"
---

# Healthcare Research Workflow

## Steps
1. Parse query into structured terms (disease, intervention, outcome)
2. Search PubMed and clinicaltrials.gov in parallel
3. For each result: extract claim, title, URL, date, journal
4. Label evidence tier:
   - 🟢 RCT or meta-analysis
   - 🟡 Observational study
   - 🟠 Case report
   - 🔴 Speculation / commentary
5. Deduplicate by URL; sort newest first
6. Return JSON only, no narrative text

## Output schema
{
  "query": "<original query>",
  "claims": [
    {
      "claim": "<one-sentence finding>",
      "sources": [{"title": "...", "url": "...", "date": "YYYY-MM-DD", "tier": "🟢"}]
    }
  ]
}

## Don't
- Scrape paywalled content
- Mix marketing claims with clinical claims
- Omit dates or URLs
YAML frontmatter is parsed by Claude Code's routing. The description drives auto-match; allowed-tools enforces safety; context: fork isolates the verbose work.
.claude/skills/ci-review.md (headless mode)yaml
---
name: ci-review
description: |
  Async PR code review for CI/CD pipelines. Reads diff, flags
  regressions, suggests refactors, emits structured JSON.
  Designed for headless invocation: claude -p --skill ci-review.
context: fork
allowed-tools: [Read, Glob, Grep, Bash]
argument-hint: "PR number or branch name"
---

# CI Review

Invoked from .github/workflows/pr-review.yml as:
  claude -p --skill ci-review "PR #${{ github.event.pull_request.number }}"

## Checklist
- [ ] No new console.error or unhandled rejections
- [ ] Type signatures on new functions
- [ ] Coverage did not decrease
- [ ] No secrets in committed files
- [ ] Migrations are idempotent

## Output (single JSON object on stdout)
{
  "pr_number": 42,
  "verdict": "approved" | "needs-rework" | "hold",
  "summary": "<two-sentence overview>",
  "findings": [
    {
      "file": "<path>",
      "line": <int>,
      "severity": "error" | "warning" | "info",
      "message": "<what's wrong>",
      "suggestion": "<concrete fix>"
    }
  ]
}

## Bash usage
- Read-only commands only: `git diff`, `git log`, `grep`, `ls`
- Never run tests, linters, or anything that mutates state
Headless skills emit a single JSON object to stdout. CI parses it into a PR comment. Tool restrictions prevent the skill from accidentally mutating the repo.
06 · Distractor patterns

Looks right, isn't

Each row pairs a plausible-looking pattern with the failure it actually creates. These are the shapes exam distractors are built from.

Looks right

Drop a markdown file in .claude/skills/ and Claude will pick it up.

Actually wrong

Skills require YAML frontmatter (name, description, optional allowed-tools / context / argument-hint). Without it, the file is plain markdown and won't auto-match queries.

Looks right

Build one large mega-skill that covers all team workflows.

Actually wrong

Skills should be granular, one workflow per skill. Mega-skills break context isolation and make tool restrictions too broad. Aim for skills under ~300 lines.

Looks right

Use a Skill instead of CLAUDE.md so it doesn't load on every session.

Actually wrong

Skills and CLAUDE.md serve different purposes. CLAUDE.md = always-on team standards. Skills = on-demand workflows triggered by description match. Use both, not one.

07 · Compare

Side-by-side

AspectSkillCLAUDE.mdSlash command (legacy)
LoadingOn-demand (description match)Every sessionOn explicit invocation
Tool restrictionYes (allowed-tools)NoNo
Output isolationYes (context: fork)No (in-session)No
Best forReusable workflowsAlways-on standardsQuick aliases (deprecating)
08 · When to use

Decision tree

01

Is this a reusable workflow you'll invoke many times?

YesMake it a Skill in .claude/skills/. Write a description that matches likely user phrasings.
NoKeep it inline or in CLAUDE.md if it's a one-time rule.
02

Should the workflow run with restricted tools (e.g., read-only)?

YesSet allowed-tools in the frontmatter. Skills enforce the restriction; CLAUDE.md cannot.
NoSkill or CLAUDE.md both work. Pick by load pattern.
03

Should the verbose work stay out of the main conversation?

YesAdd context: fork to the frontmatter. Skill output is summarized; intermediate work is hidden.
NoSkip fork; let work flow into the main context.
09 · On the exam

Question patterns

Skills exam trap — painterly cautionary scene featuring Loop mascot.
Your skill description is "helps with code" and Claude activates it for every request. Fix?
Description is too generic. Rewrite as task-specific: "validates Python security patterns; use when reviewing code for SQL injection or hardcoded secrets." Specific descriptions enable accurate matching.
A skill has `allowed-tools: Read, Grep, Glob, Bash` and you ask Claude to edit a file. It refuses. Why?
Skills enforce allowed-tools at activation time. While the skill is active, Claude can only use those tools. Either add Edit to allowed-tools, or invoke a different skill that has Edit.
You wrote a 2,000-line SKILL.md. Claude responses are slow when the skill is active. Better structure?
Use progressive disclosure. Keep SKILL.md under 500 lines. Move detailed references to references/, examples to examples/, scripts to scripts/. Link from SKILL.md: "if user asks about X, read references/guide.md."
Your skill works on your machine but a teammate can't trigger it. Where's the file?
Personal skills (~/.claude/skills/) don't sync. Move it to .claude/skills/ (project, version-controlled). Commit and push. Teammates pull and the skill is available.
Two skills match the same request. Which one activates?
Both load into context simultaneously. If their instructions conflict, results are unpredictable. Make descriptions non-overlapping: each skill should describe a unique task with no semantic overlap.
A skill has scripts in `scripts/` but they don't execute when invoked. Why?
Scripts must be marked executable (chmod +x) and the SKILL.md must reference them with the correct relative path. Test the script standalone first; then verify SKILL.md instructions are clear about when/how to run it.
When does a skill trump CLAUDE.md, and when does CLAUDE.md trump a skill?
CLAUDE.md is always-loaded, applying to every conversation. Skills load on-demand when their description matches. They're complementary: CLAUDE.md for universal team standards; skills for repeated task workflows. Neither trumps the other; they layer.
Your CI pipeline invokes Claude with `--skill ci-review` but the skill isn't activated. What's wrong?
The flag must match the skill's name field exactly (lowercase, hyphenated). Check the SKILL.md frontmatter. Also verify the skill file is in .claude/skills/ (not ~/.claude/skills/ for project-scoped CI).
10 · FAQ

Frequently asked

What does context: fork do?
Runs the skill in an isolated subagent and returns a summary. Keeps the main conversation clean.
Can I invoke a skill explicitly?
Yes, type @skill skill-name or just the name. Most flows rely on auto-detection via description.
Do subagents inherit my skills?
No, list them in the subagent's frontmatter skills field if you want them available.