writing-skills
Use when creating new skills, editing existing skills, or verifying skills work before deployment
| Model | Source | Category |
|---|---|---|
| opus | core | Meta |
Tools: Read, Glob, Grep, Bash, Edit, Write, Task, WebSearch, WebFetch
Context: fork
Overview
Section titled “Overview”Mandatory Announcement — FIRST OUTPUT before anything else:
┏━ 🛡 writing-skills ━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓┃ [one-line description of what skill you're writing] ┃┗━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┛No exceptions. Box frame first, then work.
What is this? Core principle in 1-2 sentences.
Full Reference
Writing Skills
Section titled “Writing Skills”Overview
Section titled “Overview”Mandatory Announcement — FIRST OUTPUT before anything else:
┏━ 🛡 writing-skills ━━━━━━━━━━━━━━━━━━━━━━━━━━━━┓┃ [one-line description of what skill you're writing] ┃┗━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━┛No exceptions. Box frame first, then work.
Skill Discipline
Section titled “Skill Discipline”Three non-negotiable constraints for every skill:
1. Size Limit: 500 Lines Max
Section titled “1. Size Limit: 500 Lines Max”SKILL.md files MUST be under 500 lines. If your skill content exceeds this:
- Move detailed reference material to
reference/*.mdfiles - Keep SKILL.md as a lean entry point (<100 lines for reference skills, <300 for process skills)
- The 500-line limit exists because large skill files waste context tokens
2. Provenance Fields
Section titled “2. Provenance Fields”Every SKILL.md frontmatter MUST include:
---name: skill-namedescription: Third-person description...model: claude-sonnet-4-6last_updated: 2026-03-01category: Category---The last_updated field is automatically maintained by the update-last-updated hook.
3. Third-Person Descriptions
Section titled “3. Third-Person Descriptions”Skill descriptions MUST be written in third person:
- ✓ “Processes video files and generates thumbnails”
- ✓ “Routes frontend requests to the correct sub-skill”
- ✗ “I can help you process video files”
- ✗ “Use me when you need to process videos”
Why: First-person descriptions cause inconsistent discovery behavior. Agents pattern-match differently on “I can” vs “Processes X”.
Writing skills IS Test-Driven Development applied to process documentation.
Personal skills live in agent-specific directories (~/.claude/skills for Claude Code, ~/.agents/skills/ for Codex)
You write test cases (pressure scenarios with subagents), watch them fail (baseline behavior), write the skill (documentation), watch tests pass (agents comply), and refactor (close loopholes).
Core principle: If you didn’t watch an agent fail without the skill, you don’t know if the skill teaches the right thing.
REQUIRED BACKGROUND: You MUST understand armadillo:test-driven-development before using this skill. That skill defines the fundamental RED-GREEN-REFACTOR cycle. This skill adapts TDD to documentation.
Official guidance: For Anthropic’s official skill authoring best practices, see anthropic-best-practices.md. This document provides additional patterns and guidelines that complement the TDD-focused approach in this skill.
Model requirement: Skill authoring requires understanding persuasion principles and constraint design. Use Opus 4.6 (claude-opus-4-6).
What is a Skill?
Section titled “What is a Skill?”A skill is a reference guide for proven techniques, patterns, or tools. Skills help future Claude instances find and apply effective approaches.
Skills are: Reusable techniques, patterns, tools, reference guides
Skills are NOT: Narratives about how you solved a problem once
TDD Mapping for Skills
Section titled “TDD Mapping for Skills”| TDD Concept | Skill Creation |
|---|---|
| Test case | Pressure scenario with subagent |
| Production code | Skill document (SKILL.md) |
| Test fails (RED) | Agent violates rule without skill (baseline) |
| Test passes (GREEN) | Agent complies with skill present |
| Refactor | Close loopholes while maintaining compliance |
| Write test first | Run baseline scenario BEFORE writing skill |
| Watch it fail | Document exact rationalizations agent uses |
| Minimal code | Write skill addressing those specific violations |
| Watch it pass | Verify agent now complies |
| Refactor cycle | Find new rationalizations → plug → re-verify |
The entire skill creation process follows RED-GREEN-REFACTOR.
When to Create a Skill
Section titled “When to Create a Skill”Create when:
- Technique wasn’t intuitively obvious to you
- You’d reference this again across projects
- Pattern applies broadly (not project-specific)
- Others would benefit
Don’t create for:
- One-off solutions
- Standard practices well-documented elsewhere
- Project-specific conventions (put in CLAUDE.md)
- Mechanical constraints (if it’s enforceable with regex/validation, automate it—save documentation for judgment calls)
Skill Types
Section titled “Skill Types”Technique
Section titled “Technique”Concrete method with steps to follow (condition-based-waiting, root-cause-tracing)
Pattern
Section titled “Pattern”Way of thinking about problems (flatten-with-flags, test-invariants)
Reference
Section titled “Reference”API docs, syntax guides, tool documentation (office docs)
Directory Structure
Section titled “Directory Structure”skills/ skill-name/ SKILL.md # Main reference (required) supporting-file.* # Only if neededFlat namespace - all skills in one searchable namespace
Separate files for:
- Heavy reference (100+ lines) - API docs, comprehensive syntax
- Reusable tools - Scripts, utilities, templates
Keep inline:
- Principles and concepts
- Code patterns (< 50 lines)
- Everything else
Living Reference Pattern
Section titled “Living Reference Pattern”When a skill has a reference/ directory, SKILL.md must be a lean router — not an inline knowledge dump.
Rule: If reference docs exist, SKILL.md must NOT contain inline API facts, code examples teaching the technology, configuration details, or version-specific gotchas. Those belong in reference/*.md files.
Lean Router Contents
Section titled “Lean Router Contents”| Section | Required | Notes |
|---|---|---|
| Frontmatter | ✓ | name, description only |
| Overview | ✓ | 1-3 sentences — what + when |
| Quick Reference | ✓ | Decision table for common operations |
| Reference Index | ✓ | Table pointing to reference/*.md files |
| Breaking Changes | ○ | Only when version-sensitive facts exist |
Reference Index Table Format
Section titled “Reference Index Table Format”## Reference Index
| File | Contents ||------|----------|| `reference/api.md` | Endpoints, request/response shapes, auth || `reference/config.md` | Configuration options, environment vars || `reference/errors.md` | Error codes, troubleshooting |What Stays in SKILL.md vs reference/
Section titled “What Stays in SKILL.md vs reference/”| Belongs in SKILL.md | Belongs in reference/ |
|---|---|
| When to use this skill | API endpoint specs |
| Which reference file to check | Code examples (>10 lines) |
| Decision logic | Configuration schemas |
| Common operation names | Version-specific gotchas |
Self-contained skills (no reference/ directory) are exempt — keep everything inline.
SKILL.md Structure
Section titled “SKILL.md Structure”Frontmatter (YAML):
- Only two fields supported:
nameanddescription - Max 1024 characters total
name: Use letters, numbers, and hyphens only (no parentheses, special chars)description: Third-person, describes WHAT the skill does and WHEN to use it (never HOW)- Include both purpose (what it does) and triggering conditions (when to use it)
- Start with a brief statement of what it does, then “Use when…” for triggers
- NEVER summarize the skill’s process or workflow (see CSO section for why)
- Keep under 500 characters if possible
---name: Skill-Name-With-Hyphensdescription: Use when [specific triggering conditions and symptoms]---
# Skill Name
## OverviewWhat is this? Core principle in 1-2 sentences.
## When to Use[Small inline flowchart IF decision non-obvious]
Bullet list with SYMPTOMS and use casesWhen NOT to use
## Core Pattern (for techniques/patterns)Before/after code comparison
## Quick ReferenceTable or bullets for scanning common operations
## ImplementationInline code for simple patternsLink to file for heavy reference or reusable tools
## Common MistakesWhat goes wrong + fixes
## Real-World Impact (optional)Concrete resultsClaude Search Optimization (CSO)
Section titled “Claude Search Optimization (CSO)”Critical for discovery: Future Claude needs to FIND your skill
1. Rich Description Field
Section titled “1. Rich Description Field”Purpose: Claude reads description to decide which skills to load for a given task. Make it answer: “Should I read this skill right now?”
Formula: WHAT + WHEN, never HOW
Include both what the skill does (purpose) and when to use it (triggers). Never describe the skill’s process, workflow, or steps.
CRITICAL: Description = What It Does + When to Use, NEVER How It Works
A good description tells Claude the skill’s purpose and the situations that trigger it. A bad description summarizes the skill’s workflow or process steps.
Why “never HOW” matters: Testing revealed that when a description summarizes the skill’s workflow, Claude may follow the description instead of reading the full skill content. A description saying “code review between tasks” caused Claude to do ONE review, even though the skill’s flowchart clearly showed TWO reviews (spec compliance then code quality).
When the description included purpose + triggers but no workflow summary, Claude correctly read the flowchart and followed the two-stage review process.
The trap: Descriptions that summarize workflow create a shortcut Claude will take. The skill body becomes documentation Claude skips.
# ❌ BAD: Summarizes workflow (HOW) - Claude may follow this instead of reading skilldescription: Use when executing plans - dispatches subagent per task with code review between tasks
# ❌ BAD: Too much process detail (HOW)description: Use for TDD - write test first, watch it fail, write minimal code, refactor
# ✅ GOOD: WHAT + WHEN, no workflow summarydescription: Dispatch fresh subagent per task with two-stage review. Use when executing implementation plans with independent tasks in the current session
# ✅ GOOD: WHAT + WHENdescription: Enforce test-first development discipline. Use when implementing any feature or bugfix, before writing implementation codeContent:
- Start with a brief statement of what the skill does (WHAT)
- Follow with concrete triggers, symptoms, and situations (WHEN)
- Describe the problem (race conditions, inconsistent behavior) not language-specific symptoms (setTimeout, sleep)
- Keep triggers technology-agnostic unless the skill itself is technology-specific
- If skill is technology-specific, make that explicit in the trigger
- Write in third person (injected into system prompt)
- NEVER summarize the skill’s process, steps, or workflow (HOW)
# ❌ BAD: Too abstract, vague, doesn't include when to usedescription: For async testing
# ❌ BAD: First persondescription: I can help you with async tests when they're flaky
# ❌ BAD: Mentions technology but skill isn't specific to itdescription: Use when tests use setTimeout/sleep and are flaky
# ✅ GOOD: Starts with "Use when", describes problem, no workflowdescription: Use when tests have race conditions, timing dependencies, or pass/fail inconsistently
# ✅ GOOD: Technology-specific skill with explicit triggerdescription: Use when using React Router and handling authentication redirects2. Keyword Coverage
Section titled “2. Keyword Coverage”Use words Claude would search for:
- Error messages: “Hook timed out”, “ENOTEMPTY”, “race condition”
- Symptoms: “flaky”, “hanging”, “zombie”, “pollution”
- Synonyms: “timeout/hang/freeze”, “cleanup/teardown/afterEach”
- Tools: Actual commands, library names, file types
3. Descriptive Naming
Section titled “3. Descriptive Naming”Use active voice, verb-first:
- ✅
creating-skillsnotskill-creation - ✅
condition-based-waitingnotasync-test-helpers
4. Token Efficiency (Critical)
Section titled “4. Token Efficiency (Critical)”Problem: getting-started and frequently-referenced skills load into EVERY conversation. Every token counts.
Target word counts:
- getting-started workflows: <150 words each
- Frequently-loaded skills: <200 words total
- Other skills: <500 words (still be concise)
Techniques:
Move details to tool help:
# ❌ BAD: Document all flags in SKILL.mdsearch-conversations supports --text, --both, --after DATE, --before DATE, --limit N
# ✅ GOOD: Reference --helpsearch-conversations supports multiple modes and filters. Run --help for details.Use cross-references:
# ❌ BAD: Repeat workflow detailsWhen searching, dispatch subagent with template...[20 lines of repeated instructions]
# ✅ GOOD: Reference other skillAlways use subagents (50-100x context savings). REQUIRED: Use [other-skill-name] for workflow.Compress examples:
# ❌ BAD: Verbose example (42 words)your human partner: "How did we handle authentication errors in React Router before?"You: I'll search past conversations for React Router authentication patterns.[Dispatch subagent with search query: "React Router authentication error handling 401"]
# ✅ GOOD: Minimal example (20 words)Partner: "How did we handle auth errors in React Router?"You: Searching...[Dispatch subagent → synthesis]Eliminate redundancy:
- Don’t repeat what’s in cross-referenced skills
- Don’t explain what’s obvious from command
- Don’t include multiple examples of same pattern
Verification:
wc -w skills/path/SKILL.md# getting-started workflows: aim for <150 each# Other frequently-loaded: aim for <200 totalName by what you DO or core insight:
- ✅
condition-based-waiting>async-test-helpers - ✅
using-skillsnotskill-usage - ✅
flatten-with-flags>data-structure-refactoring - ✅
root-cause-tracing>debugging-techniques
Gerunds (-ing) work well for processes:
creating-skills,testing-skills,debugging-with-logs- Active, describes the action you’re taking
4. Cross-Referencing Other Skills
Section titled “4. Cross-Referencing Other Skills”When writing documentation that references other skills:
Use skill name only, with explicit requirement markers:
- ✅ Good:
**REQUIRED SUB-SKILL:** Use armadillo:test-driven-development - ✅ Good:
**REQUIRED BACKGROUND:** You MUST understand armadillo:systematic-debugging - ❌ Bad:
See skills/testing/test-driven-development(unclear if required) - ❌ Bad:
@skills/testing/test-driven-development/SKILL.md(force-loads, burns context)
Why no @ links: @ syntax force-loads files immediately, consuming 200k+ context before you need them.
Flowchart Usage
Section titled “Flowchart Usage”digraph when_flowchart { "Need to show information?" [shape=diamond]; "Decision where I might go wrong?" [shape=diamond]; "Use markdown" [shape=box]; "Small inline flowchart" [shape=box];
"Need to show information?" -> "Decision where I might go wrong?" [label="yes"]; "Decision where I might go wrong?" -> "Small inline flowchart" [label="yes"]; "Decision where I might go wrong?" -> "Use markdown" [label="no"];}Use flowcharts ONLY for:
- Non-obvious decision points
- Process loops where you might stop too early
- “When to use A vs B” decisions
Never use flowcharts for:
- Reference material → Tables, lists
- Code examples → Markdown blocks
- Linear instructions → Numbered lists
- Labels without semantic meaning (step1, helper2)
See @graphviz-conventions.dot for graphviz style rules.
Visualizing for your human partner: Use render-graphs.js in this directory to render a skill’s flowcharts to SVG:
./render-graphs.js ../some-skill # Each diagram separately./render-graphs.js ../some-skill --combine # All diagrams in one SVGCode Examples
Section titled “Code Examples”One excellent example beats many mediocre ones
Choose most relevant language:
- Testing techniques → TypeScript/JavaScript
- System debugging → Shell/Python
- Data processing → Python
Good example:
- Complete and runnable
- Well-commented explaining WHY
- From real scenario
- Shows pattern clearly
- Ready to adapt (not generic template)
Don’t:
- Implement in 5+ languages
- Create fill-in-the-blank templates
- Write contrived examples
You’re good at porting - one great example is enough.
File Organization
Section titled “File Organization”Self-Contained Skill
Section titled “Self-Contained Skill”defense-in-depth/ SKILL.md # Everything inlineWhen: All content fits, no heavy reference needed
Skill with Reusable Tool
Section titled “Skill with Reusable Tool”condition-based-waiting/ SKILL.md # Overview + patterns example.ts # Working helpers to adaptWhen: Tool is reusable code, not just narrative
Skill with Heavy Reference
Section titled “Skill with Heavy Reference”pptx/ SKILL.md # Overview + workflows pptxgenjs.md # 600 lines API reference ooxml.md # 500 lines XML structure scripts/ # Executable toolsWhen: Reference material too large for inline
The Iron Law (Same as TDD)
Section titled “The Iron Law (Same as TDD)”NO SKILL WITHOUT A FAILING TEST FIRSTThis applies to NEW skills AND EDITS to existing skills.
Write skill before testing? Delete it. Start over. Edit skill without testing? Same violation.
No exceptions:
- Not for “simple additions”
- Not for “just adding a section”
- Not for “documentation updates”
- Don’t keep untested changes as “reference”
- Don’t “adapt” while running tests
- Delete means delete
REQUIRED BACKGROUND: The armadillo:test-driven-development skill explains why this matters. Same principles apply to documentation.
Testing All Skill Types
Section titled “Testing All Skill Types”Different skill types need different test approaches:
Discipline-Enforcing Skills (rules/requirements)
Section titled “Discipline-Enforcing Skills (rules/requirements)”Examples: TDD, verification-before-completion, designing-before-coding
Test with:
- Academic questions: Do they understand the rules?
- Pressure scenarios: Do they comply under stress?
- Multiple pressures combined: time + sunk cost + exhaustion
- Identify rationalizations and add explicit counters
Success criteria: Agent follows rule under maximum pressure
Technique Skills (how-to guides)
Section titled “Technique Skills (how-to guides)”Examples: condition-based-waiting, root-cause-tracing, defensive-programming
Test with:
- Application scenarios: Can they apply the technique correctly?
- Variation scenarios: Do they handle edge cases?
- Missing information tests: Do instructions have gaps?
Success criteria: Agent successfully applies technique to new scenario
Pattern Skills (mental models)
Section titled “Pattern Skills (mental models)”Examples: reducing-complexity, information-hiding concepts
Test with:
- Recognition scenarios: Do they recognize when pattern applies?
- Application scenarios: Can they use the mental model?
- Counter-examples: Do they know when NOT to apply?
Success criteria: Agent correctly identifies when/how to apply pattern
Reference Skills (documentation/APIs)
Section titled “Reference Skills (documentation/APIs)”Examples: API documentation, command references, library guides
Test with:
- Retrieval scenarios: Can they find the right information?
- Application scenarios: Can they use what they found correctly?
- Gap testing: Are common use cases covered?
Success criteria: Agent finds and correctly applies reference information
Common Rationalizations for Skipping Testing
Section titled “Common Rationalizations for Skipping Testing”| Excuse | Reality |
|---|---|
| ”Skill is obviously clear” | Clear to you ≠ clear to other agents. Test it. |
| ”It’s just a reference” | References can have gaps, unclear sections. Test retrieval. |
| ”Testing is overkill” | Untested skills have issues. Always. 15 min testing saves hours. |
| ”I’ll test if problems emerge” | Problems = agents can’t use skill. Test BEFORE deploying. |
| ”Too tedious to test” | Testing is less tedious than debugging bad skill in production. |
| ”I’m confident it’s good” | Overconfidence guarantees issues. Test anyway. |
| ”Academic review is enough” | Reading ≠ using. Test application scenarios. |
| ”No time to test” | Deploying untested skill wastes more time fixing it later. |
All of these mean: Test before deploying. No exceptions.
Bulletproofing Skills Against Rationalization
Section titled “Bulletproofing Skills Against Rationalization”Skills that enforce discipline (like TDD) need to resist rationalization. Agents are smart and will find loopholes when under pressure.
Psychology note: Understanding WHY persuasion techniques work helps you apply them systematically. See persuasion-principles.md for research foundation (Cialdini, 2021; Meincke et al., 2025) on authority, commitment, scarcity, social proof, and unity principles.
Close Every Loophole Explicitly
Section titled “Close Every Loophole Explicitly”Don’t just state the rule - forbid specific workarounds:
No exceptions:
- Don’t keep it as “reference”
- Don’t “adapt” it while writing tests
- Don’t look at it
- Delete means delete
</Good>
### Address "Spirit vs Letter" Arguments
Add foundational principle early:
```markdown**Violating the letter of the rules is violating the spirit of the rules.**This cuts off entire class of “I’m following the spirit” rationalizations.
Build Rationalization Table
Section titled “Build Rationalization Table”Capture rationalizations from baseline testing (see Testing section below). Every excuse agents make goes in the table:
| Excuse | Reality ||--------|---------|| "Too simple to test" | Simple code breaks. Test takes 30 seconds. || "I'll test after" | Tests passing immediately prove nothing. || "Tests after achieve same goals" | Tests-after = "what does this do?" Tests-first = "what should this do?" |Create Red Flags List
Section titled “Create Red Flags List”Make it easy for agents to self-check when rationalizing:
## Red Flags - STOP and Start Over
- Code before test- "I already manually tested it"- "Tests after achieve the same purpose"- "It's about spirit not ritual"- "This is different because..."
**All of these mean: Delete code. Start over with TDD.**Update CSO for Violation Symptoms
Section titled “Update CSO for Violation Symptoms”Add to description: symptoms of when you’re ABOUT to violate the rule:
description: use when implementing any feature or bugfix, before writing implementation codeRED-GREEN-REFACTOR for Skills
Section titled “RED-GREEN-REFACTOR for Skills”Follow the TDD cycle:
RED: Write Failing Test (Baseline)
Section titled “RED: Write Failing Test (Baseline)”Run pressure scenario with subagent WITHOUT the skill. Document exact behavior:
- What choices did they make?
- What rationalizations did they use (verbatim)?
- Which pressures triggered violations?
This is “watch the test fail” - you must see what agents naturally do before writing the skill.
GREEN: Write Minimal Skill
Section titled “GREEN: Write Minimal Skill”Write skill that addresses those specific rationalizations. Don’t add extra content for hypothetical cases.
Run same scenarios WITH skill. Agent should now comply.
REFACTOR: Close Loopholes
Section titled “REFACTOR: Close Loopholes”Agent found new rationalization? Add explicit counter. Re-test until bulletproof.
Testing methodology: See @testing-skills-with-subagents.md for the complete testing methodology:
- How to write pressure scenarios
- Pressure types (time, sunk cost, authority, exhaustion)
- Plugging holes systematically
- Meta-testing techniques
Anti-Patterns
Section titled “Anti-Patterns”❌ Narrative Example
Section titled “❌ Narrative Example”“In session 2025-10-03, we found empty projectDir caused…” Why bad: Too specific, not reusable
❌ Multi-Language Dilution
Section titled “❌ Multi-Language Dilution”example-js.js, example-py.py, example-go.go Why bad: Mediocre quality, maintenance burden
❌ Code in Flowcharts
Section titled “❌ Code in Flowcharts”step1 [label="import fs"];step2 [label="read file"];Why bad: Can’t copy-paste, hard to read
❌ Generic Labels
Section titled “❌ Generic Labels”helper1, helper2, step3, pattern4 Why bad: Labels should have semantic meaning
STOP: Before Moving to Next Skill
Section titled “STOP: Before Moving to Next Skill”After writing ANY skill, you MUST STOP and complete the deployment process.
Do NOT:
- Create multiple skills in batch without testing each
- Move to next skill before current one is verified
- Skip testing because “batching is more efficient”
The deployment checklist below is MANDATORY for EACH skill.
Deploying untested skills = deploying untested code. It’s a violation of quality standards.
Skill Creation Checklist (TDD Adapted)
Section titled “Skill Creation Checklist (TDD Adapted)”IMPORTANT: Use TodoWrite to create todos for EACH checklist item below.
RED Phase - Write Failing Test:
- Create pressure scenarios (3+ combined pressures for discipline skills)
- Run scenarios WITHOUT skill - document baseline behavior verbatim
- Identify patterns in rationalizations/failures
GREEN Phase - Write Minimal Skill:
- Name uses only letters, numbers, hyphens (no parentheses/special chars)
- YAML frontmatter with only name and description (max 1024 chars)
- Description starts with “Use when…” and includes specific triggers/symptoms
- Description written in third person
- Keywords throughout for search (errors, symptoms, tools)
- Clear overview with core principle
- Address specific baseline failures identified in RED
- Code inline OR link to separate file
- One excellent example (not multi-language)
- Run scenarios WITH skill - verify agents now comply
REFACTOR Phase - Close Loopholes:
- Identify NEW rationalizations from testing
- Add explicit counters (if discipline skill)
- Build rationalization table from all test iterations
- Create red flags list
- Re-test until bulletproof
Quality Checks:
- Small flowchart only if decision non-obvious
- Quick reference table
- Common mistakes section
- No narrative storytelling
- Supporting files only for tools or heavy reference
Deployment:
- Commit skill to git and push to your fork (if configured)
- Consider contributing back via PR (if broadly useful)
Discovery Workflow
Section titled “Discovery Workflow”How future Claude finds your skill:
- Encounters problem (“tests are flaky”)
- Finds SKILL (description matches)
- Scans overview (is this relevant?)
- Reads patterns (quick reference table)
- Loads example (only when implementing)
Optimize for this flow - put searchable terms early and often.
The Bottom Line
Section titled “The Bottom Line”Creating skills IS TDD for process documentation.
Same Iron Law: No skill without failing test first. Same cycle: RED (baseline) → GREEN (write skill) → REFACTOR (close loopholes). Same benefits: Better quality, fewer surprises, bulletproof results.
If you follow TDD for code, follow it for skills. It’s the same discipline applied to documentation.