What do practitioners actually DO (not just document)? Methodology for extracting practitioner knowledge from YouTube video transcripts. Use when (1) researching how people actually use a technology, (2) seeking practitioner insights beyond documentation, (3) looking for tips, patterns, or workflows from creators, (4) building a research corpus from video content.
Seed question: What do practitioners actually DO (not just document)?
Extract structured knowledge from YouTube video transcripts. Videos capture practitioner insights, tips, and patterns that don't appear in formal documentation.
Method A: Web Search
WebSearch: "[topic] tutorial 2025 youtube"
WebSearch: "[topic] best practices youtube"
WebSearch: "[topic] tips advanced youtube"
Method B: Playwright Discovery (Better for Recommendations)
[topic] tutorial [year]Method C: Known Creator Channels
Delegates to video-transcript-extraction skill.
For each discovered video:
For each transcript, extract:
Structural Elements:
Quality Signals:
Cross-Video Analysis:
Output Structure:
# [Topic] YouTube Research Analysis
**Date:** YYYY-MM-DD
**Videos Analyzed:** N
## Videos Analyzed
| Video | Creator | Focus | Key Value |
## New Patterns Discovered
## Reinforced Patterns
## Contradictions Found
## Methodology Notes
Normal mode:
output/
HANDOFF.md — Summary + suggested next actions
corpus.json — Machine-readable (auto-generated)
all_content.md — Consolidated markdown
analysis/
overview.md
patterns.md
themes.md
Budget mode: Skip corpus.json unless user requests. HANDOFF.md and all_content.md always generated.
Relational memory: Memorize key findings if relational-memory MCP is configured (skip if not available):
mcp__relational-memory__memorize(
agent_name="youtube-research",
layer="recent",
content="[key finding]",
metadata={"topic": "...", "videos": N}
)
Choose detail level BEFORE starting extraction:
| Level | Mode | Per-Video Output | When to Use |
|---|---|---|---|
| 0-3 | Quick | 1-2 sentences, topic tags | Triage many videos, initial discovery |
| 4-6 | Balanced | Summary + key points + notable quotes | Standard research, known-good sources |
| 7-10 | Deep | Full extraction, timestamps, cross-references | High-value topics, building corpus |
Adjusting mid-research: Start at 4-6 by default. Increase to 7-10 if finding gold. Decrease to 0-3 if hitting diminishing returns.
This skill is SELF-MANAGING. No user input needed for iteration decisions.
## SELF-REVIEW - Pass N
### Value Assessment
1. Patterns found this pass: [count]
2. Novel insights (not seen before): [count]
3. Reinforced patterns: [count]
### Budget Check
4. Detail level used: [N]
5. Token investment: HIGH/MEDIUM/LOW
6. Value delivered: HIGH/MEDIUM/LOW
7. Value/Token ratio: GOOD/ACCEPTABLE/POOR
### Decisions
8. Continue? [YES/NO]
9. Adjust detail level? [UP/DOWN/SAME]
def should_continue():
consecutive_low = count_trailing_lows(pass_history)
if consecutive_low >= 2:
return STOP, "Data exhausted"
if saturation == "YES":
return STOP, "Saturation confirmed"
if value_token_ratio == "POOR" and consecutive_low >= 1:
return STOP, "Diminishing returns"
return CONTINUE, "Proceed to next pass"
Consult reference/topic-based-escalation.md when extracted content touches:
Mid-research escalation is a suggestion, not automatic. Note it and let the user decide.
Activation (any of these):
--budget or -bbudget-mode skill is active in sessionWhen active:
Note: After context compaction, auto-detection may fail. Re-invoke budget-mode skill or pass --budget explicitly.
Propagation: When invoking other skills, pass budget: "Invoking video-transcript-extraction --budget"
A vasana is a pattern that persists across unrelated contexts. If during
this task you notice such a pattern emerging, it may be worth capturing.
This skill works best alongside the vasana skill and vasana hook
from the Vasana System plugin.
Modify freely. Keep this section intact.