From research-toolkit
What do practitioners actually DO (not just document)? Methodology for extracting practitioner knowledge from YouTube video transcripts. Use when (1) researching how people actually use a technology, (2) seeking practitioner insights beyond documentation, (3) looking for tips, patterns, or workflows from creators, (4) building a research corpus from video content.
npx claudepluginhub bogheorghiu/ex-cog-dev --plugin research-toolkitThis skill uses the workspace's default tool permissions.
**Seed question:** *What do practitioners actually DO (not just document)?*
Provides UI/UX resources: 50+ styles, color palettes, font pairings, guidelines, charts for web/mobile across React, Next.js, Vue, Svelte, Tailwind, React Native, Flutter. Aids planning, building, reviewing interfaces.
Fetches up-to-date documentation from Context7 for libraries and frameworks like React, Next.js, Prisma. Use for setup questions, API references, and code examples.
Applies Acme Corporation brand guidelines including colors, fonts, layouts, and messaging to generated PowerPoint, Excel, and PDF documents.
Share bugs, ideas, or general feedback.
Seed question: What do practitioners actually DO (not just document)?
Extract structured knowledge from YouTube video transcripts. Videos capture practitioner insights, tips, and patterns that don't appear in formal documentation.
Method A: Web Search
WebSearch: "[topic] tutorial 2025 youtube"
WebSearch: "[topic] best practices youtube"
WebSearch: "[topic] tips advanced youtube"
Method B: Playwright Discovery (Better for Recommendations)
[topic] tutorial [year]Method C: Known Creator Channels
Delegates to video-transcript-extraction skill.
For each discovered video:
For each transcript, extract:
Structural Elements:
Quality Signals:
Cross-Video Analysis:
Output Structure:
# [Topic] YouTube Research Analysis
**Date:** YYYY-MM-DD
**Videos Analyzed:** N
## Videos Analyzed
| Video | Creator | Focus | Key Value |
## New Patterns Discovered
## Reinforced Patterns
## Contradictions Found
## Methodology Notes
Normal mode:
output/
HANDOFF.md — Summary + suggested next actions
corpus.json — Machine-readable (auto-generated)
all_content.md — Consolidated markdown
analysis/
overview.md
patterns.md
themes.md
Budget mode: Skip corpus.json unless user requests. HANDOFF.md and all_content.md always generated.
Relational memory: Memorize key findings if relational-memory MCP is configured (skip if not available):
mcp__relational-memory__memorize(
agent_name="youtube-research",
layer="recent",
content="[key finding]",
metadata={"topic": "...", "videos": N}
)
Choose detail level BEFORE starting extraction:
| Level | Mode | Per-Video Output | When to Use |
|---|---|---|---|
| 0-3 | Quick | 1-2 sentences, topic tags | Triage many videos, initial discovery |
| 4-6 | Balanced | Summary + key points + notable quotes | Standard research, known-good sources |
| 7-10 | Deep | Full extraction, timestamps, cross-references | High-value topics, building corpus |
Adjusting mid-research: Start at 4-6 by default. Increase to 7-10 if finding gold. Decrease to 0-3 if hitting diminishing returns.
This skill is SELF-MANAGING. No user input needed for iteration decisions.
## SELF-REVIEW - Pass N
### Value Assessment
1. Patterns found this pass: [count]
2. Novel insights (not seen before): [count]
3. Reinforced patterns: [count]
### Budget Check
4. Detail level used: [N]
5. Token investment: HIGH/MEDIUM/LOW
6. Value delivered: HIGH/MEDIUM/LOW
7. Value/Token ratio: GOOD/ACCEPTABLE/POOR
### Decisions
8. Continue? [YES/NO]
9. Adjust detail level? [UP/DOWN/SAME]
def should_continue():
consecutive_low = count_trailing_lows(pass_history)
if consecutive_low >= 2:
return STOP, "Data exhausted"
if saturation == "YES":
return STOP, "Saturation confirmed"
if value_token_ratio == "POOR" and consecutive_low >= 1:
return STOP, "Diminishing returns"
return CONTINUE, "Proceed to next pass"
Consult reference/topic-based-escalation.md when extracted content touches:
Mid-research escalation is a suggestion, not automatic. Note it and let the user decide.
Activation (any of these):
--budget or -bbudget-mode skill is active in sessionWhen active:
Note: After context compaction, auto-detection may fail. Re-invoke budget-mode skill or pass --budget explicitly.
Propagation: When invoking other skills, pass budget: "Invoking video-transcript-extraction --budget"
A vasana is a pattern that persists across unrelated contexts. If during
this task you notice such a pattern emerging, it may be worth capturing.
This skill works best alongside the vasana skill and vasana hook
from the Vasana System plugin.
Modify freely. Keep this section intact.