Evaluate a single skill's quality against 8 completeness categories derived from Anthropic's official skills repository. Scores preparation, progression, verification, scripts, examples, anti-patterns, references, and assets. Generates scored report to .claude/audits/. Use when auditing skill quality, checking marketplace readiness, evaluating skill completeness score, performing pre-publication evaluation, or comparing to Anthropic skills.
Evaluates a single skill directory against 8 quality categories derived from Anthropic's official skills repository. Each category is scored 0-3, producing an overall completeness percentage and actionable recommendations for improvement.
Invoke this skill when:
Read the skill directory structure:
skill-path/
├── SKILL.md # Required - main skill definition
├── scripts/ # Optional - executable automation
├── references/ # Optional - supporting documentation
└── assets/ # Optional - reusable output resources
Actions:
Validation:
Run through each of the 8 categories using the detailed checklist in references/skill-completeness-checklist.md.
Quality Categories:
| Category | Evaluates | Key Indicators |
|---|---|---|
| 1. Preparation | Prerequisites met before work begins | Environment verification, input inspection, metadata extraction scripts |
| 2. Progression | Concrete steps with right level of control | Clear sequence, deterministic scripts, working examples, decision trees |
| 3. Verification | Output correctness confirmed before success | Explicit verification steps, automated checks, error-correction loops, acceptance criteria |
| 4. Scripts | Executable automation for core operations | Repetitive operations scripted, --help support, edge case handling, tested output |
| 5. Examples | Teaching through demonstration | Working code with imports, exact input→output pairs, common cases, edge case handling |
| 6. Anti-Patterns | Explicit "what NOT to do" | Known failure modes documented, bad output shown, corrections side-by-side |
| 7. References | Domain knowledge AI cannot generate | API/schema/format documentation, organized sections, linked from workflow steps |
| 8. Assets | Reusable output resources bundled | Templates, fonts, images, boilerplate the AI uses (not reads) |
Evaluation Process:
For each category:
references/skill-completeness-checklist.mdCalculate overall score and write report to .claude/audits/completeness-report-{skill-slug}.md.
Report Structure:
# Skill Completeness Report: {skill-name}
**Evaluated:** {timestamp}
**Skill Path:** {absolute-path}
## Overall Score: {percentage}% ({score}/24)
| Category | Score | Label | Findings |
|----------|-------|-------|----------|
| 1. Preparation | 2 | Adequate | Environment checks present, missing metadata extraction |
| 2. Progression | 3 | Exemplary | Clear workflow, deterministic scripts, decision tree |
| ... | ... | ... | ... |
## Category Details
### 1. Preparation (2/3 - Adequate)
**What was evaluated:**
- Environment verification before starting
- Input inspection before acting
- Metadata extraction scripts
**Evidence found:**
- ✅ Environment check at SKILL.md:45-50
- ✅ Input validation at SKILL.md:65
- ❌ No metadata extraction script in scripts/
**Recommendation:**
Add a script to extract structured metadata from inputs so the AI operates on verified data instead of assumptions.
### 2. Progression (3/3 - Exemplary)
...
## Recommendations for Improvement
1. **High Priority:** Add metadata extraction script (Preparation)
2. **Medium Priority:** Include anti-pattern examples (Anti-Patterns)
3. **Low Priority:** Add visual validation examples (Verification)
## Reference
This audit follows patterns from Anthropic's official skills repository:
- https://github.com/anthropics/skills
Checklist: [Skill Completeness Checklist](./references/skill-completeness-checklist.md)
Output Location:
Report written to .claude/audits/completeness-report-{skill-slug}.md
If .claude/audits/ does not exist, create it.
Each category is scored 0-3 based on presence and quality of evidence:
| Score | Label | Meaning | Criteria |
|---|---|---|---|
| 0 | None | Category not addressed | No evidence found for any checklist items |
| 1 | Minimal | Basic attempt, significant gaps | 1-2 checklist items present, core patterns missing |
| 2 | Adequate | Meets expectations, minor gaps | 3-4 checklist items present, core patterns followed |
| 3 | Exemplary | Exceeds expectations, Anthropic patterns | All or most checklist items present, matches Anthropic quality |
Overall Score Calculation:
Sum of category scores / 24 * 100 = percentage
Scoring Guidelines:
Preparation (0-3):
Progression (0-3):
Verification (0-3):
Scripts (0-3):
Examples (0-3):
Anti-Patterns (0-3):
References (0-3):
Assets (0-3):
Report filename: completeness-report-{skill-slug}.md
Where {skill-slug} is the skill directory name (e.g., audit-skill-completeness → completeness-report-audit-skill-completeness.md)
Report sections:
All 8 categories are detailed in references/skill-completeness-checklist.md with:
references/skill-completeness-checklist.md - detailed quality categories, checklist items, and examples from Anthropic's official skills repository