From deep-thought
Multi-framework content review with convergent synthesis. Each lens applies a named analytical framework — grounded in published source material, not persona impersonation — to the target content. Parallel dispatch, web research or user-provided references, convergent synthesis, prioritized action list. Triggers: expert panel, expert review, expert audit, panel review, multi-lens review, framework review, content audit, workshop audit, strategy review, get expert feedback, advisory review.
npx claudepluginhub ondrej-svec/heart-of-gold-toolkit --plugin deep-thoughtThis skill is limited to using the following tools:
Apply multiple analytical frameworks to content in parallel, then synthesize the findings into convergent signals, productive disagreements, and a prioritized action list.
Provides structured framework with specialized agents for research, analysis, architecture decisions, validation, synthesis, reviews, investigations, and reporting. For complex problems needing systematic exploration and persistent artifacts.
Reviews implementation plans through quality lenses (architecture, security, test coverage, code quality, standards, usability) and collaboratively iterates before implementation.
Runs parallel specialized AI reviewers for code, content, strategy, security, and academic audits using presets like code, full, security. Merges findings via consensus gate.
Share bugs, ideas, or general feedback.
Apply multiple analytical frameworks to content in parallel, then synthesize the findings into convergent signals, productive disagreements, and a prioritized action list.
Each lens is a named analytical framework (evolutionary design, learning science, practitioner reality-check, etc.) with traceable sources (books, articles, published talks). The skill applies those frameworks to the target content and cites specific concepts by name.
Each lens is NOT a persona impersonation. The skill never claims "Fowler says" or "Newport would argue." It claims "applying the evolutionary-design framework (source: Fowler, Refactoring 2nd ed.) to this content, the following pattern emerges." The framework is the lens. The author is the source. The AI is the reader applying the framework — like a careful graduate student who read the book, not like the author in the room.
This distinction matters because:
The skill supports three tiers of grounding, from highest to lowest fidelity:
The user provides the actual source texts (book chapters, blog posts, talk transcripts, papers) in a references/ directory or via --references <path>. Each lens reads its assigned reference material as primary context and applies the specific arguments from that material to the target content.
This is the closest to "what would the framework's author think" because the model is reading and applying THEIR actual words, not reconstructing them from training data. Citations point to specific sections of the provided material.
/expert-panel content/ --references docs/references/
Where docs/references/ contains files named by lens:
references/
learning-design/
newport-deep-work-ch4.md
ng-coursera-teaching-principles.md
evolutionary-design/
fowler-refactoring-smells-catalog.md
fowler-bliki-evolutionary-design.md
practitioner-reality/
orosz-ai-native-engineers-newsletter.md
Each lens reads only its own subdirectory. If the subdirectory is empty or absent, the lens falls back to Tier 2.
The skill searches the web for the framework authors' published work on the specific topic being reviewed. Each lens subagent runs WebSearch + WebFetch to find recent articles, blog posts, or talks, reads them, and extracts specific concepts to apply.
Better than training-data recall because it finds current publications. Still limited by search quality and the model's ability to faithfully represent what it read. Every finding must cite the source URL or publication name.
If both Tier 1 (no references provided) and Tier 2 (web search found nothing relevant) fail, the lens applies the framework's well-known published concepts (e.g., Fowler's code smells catalog, Meadows' 12 leverage points) from the model's training data. The lens must explicitly state: "No current published work found on this specific topic. Applying the framework's established concepts from [source name, publication year]."
This is the lowest tier. It works for frameworks with well-codified concepts (code smells, DORA metrics, leverage points) and poorly for frameworks that depend on evolving positions (a practitioner's current view on AI tools).
The skill MUST state which tier each lens operated at in the output. The user needs to know whether a finding is grounded in provided material, in a web-sourced article, or in training-data recall.
This skill MAY: read files, search the web, dispatch research subagents, read user-provided reference material, analyze content, produce structured findings. This skill MAY NOT: edit files, create code, modify the reviewed content, push changes. Claim that a specific person holds a specific opinion. Use "X says" or "X would argue" phrasing.
This is a review, not a fix. Present findings and priorities — the user decides what to act on.
| Shortcut | Why It Fails | The Cost |
|---|---|---|
| "I know this framework well enough to skip research" | You know your training data's snapshot. The source author may have revised their position, published new work, or addressed this exact topic. | Stale or fabricated attribution damages credibility |
| "Run all lenses on everything" | A 7-lens review of a README is overkill. Lenses that don't apply produce noise that buries real signal. | The user reads 7 reviews, trusts none |
| "Skip synthesis — just show each review" | Raw reviews are noise. The value is where 3+ lenses converge. | User does the synthesis themselves, defeats the purpose |
| "One lens is enough for this" | If one lens is enough, use /deep-thought:review instead. The panel exists for multi-perspective convergence. | Use the right tool |
| "Fowler would say..." | You are not Fowler. You are applying Fowler's published evolutionary-design framework. Say that. | False authority, unfalsifiable claim, potential misrepresentation |
Entry: User invoked /expert-panel with a topic, directory, or file set.
Read the project's AGENTS.md, CLAUDE.md, and README.md to understand:
Look for a references/ directory at the project root or at the path specified by --references. If found, inventory its subdirectories and match them to available lenses. Report which lenses have Tier 1 material and which will use Tier 2 (web research).
If the user specified lenses, use those. Otherwise, recommend a panel based on the content type:
| Content type | Recommended panel |
|---|---|
| Workshop / course curriculum | learning-design, evolutionary-design, practitioner-reality, behavioral-change |
| Technical documentation | evolutionary-design, practitioner-reality, tool-craft |
| Product strategy / architecture | evolutionary-design, practitioner-reality, systems-dynamics |
| Educational content (blog, talks) | learning-design, practitioner-reality, narrative-craft |
| Developer tools (CLI, SDK, API) | tool-craft, evolutionary-design, practitioner-reality |
Present the recommendation using AskUserQuestion:
Minimum: 3 lenses. Fewer than 3 cannot produce convergent findings — use /deep-thought:review instead.
Maximum: 8 lenses. More than 8 dilutes synthesis quality and burns context budget.
Exit: Target scope, lens selection, and grounding tiers confirmed.
Entry: Lenses selected, target scope confirmed.
For each selected lens, dispatch a parallel Agent subagent with this brief:
You are applying the [FRAMEWORK NAME] analytical framework to review
[TARGET DESCRIPTION].
## Grounding
[IF TIER 1]: Read the reference material at [REFERENCE PATH]. Extract
the specific concepts, frameworks, vocabulary, and evaluative criteria
from this material. Your review MUST cite specific sections or arguments
from the provided material. Do not supplement with your own knowledge of
the author — the provided text is your source of truth.
[IF TIER 2]: Search the web for published work by [SOURCE AUTHORS] on
[TOPIC AREA]. Prioritize their blog, newsletter, or personal site;
recent talks (last 2 years); and published books or articles on this
specific topic. Find at least 3 specific concepts or frameworks from
their published work that apply to this review. Cite the source URL or
publication for each.
[IF TIER 3]: Apply the framework's established published concepts:
[LIST THE WELL-KNOWN CONCEPTS]. State explicitly that no current
material was found and that you are applying established concepts from
[SOURCE, YEAR].
## Review the target content
Read: [FILE LIST]
Apply the framework's specific evaluative criteria to the content.
For each finding, cite which concept from the framework it applies and
where in the source material the concept comes from.
## Output format
## [Framework Name] lens
### Grounding (Tier [N])
- [Concept 1]: [source reference] — [how it applies to this content]
- [Concept 2]: [source reference] — [how it applies]
- [Concept 3]: [source reference] — [how it applies]
### Findings
[The review, using the framework's vocabulary. Never say "[Author] says"
or "[Author] would argue." Say "the [framework] identifies" or "applying
[concept name] from [source], this content...".]
### Top 3 changes
1. [Most impactful — with file reference]
2. [Second]
3. [Third]
All lens subagents run in parallel. Do not wait for one before dispatching the next.
If a Tier 2 subagent cannot find relevant published work: The subagent must say so honestly and fall back to Tier 3. "Web search did not surface [Author]'s published position on [topic]. Applying established [framework] concepts from [known source]." This is better than fabricating citations.
Exit: All subagent reviews collected.
Entry: N lens reviews collected from parallel subagents.
This is where the panel produces value that no single review can.
From each lens review, extract every distinct finding. Tag each with:
Group findings by theme. For each theme, count how many independent frameworks flagged it.
Where two frameworks directly contradict each other, name the tension instead of picking a winner.
Example: "The learning-design framework recommends expanding the demo to 35 minutes for deeper cognitive scaffolding (source: Newport, Deep Work, ch. 2 on attention residue). The practitioner-reality framework would cut it to keep the day realistic (source: Orosz, Pragmatic Engineer newsletter on workshop realism). The tension is depth vs. pace — the facilitator's room read should decide."
Note: the example above cites sources and frameworks, not personas. This is correct.
Produce a ranked action list ordered by:
Each action should name: what to change, which file(s), why (citing which frameworks and their specific reasoning), and estimated effort.
Exit: Synthesis complete.
Entry: Synthesis complete.
# Framework Panel Review: [target description]
## Panel composition
| Framework | Grounding | Focus |
|---|---|---|
| [Name] | Tier [N]: [source description] | [one-line focus] |
## Honesty note
This review applies named analytical frameworks to the target content.
It does not represent the personal opinions of any named author. Findings
are traceable to specific published concepts cited in each lens's grounding
section. [Tier 1 lenses] were grounded in user-provided reference material.
[Tier 2 lenses] were grounded in web-sourced publications. [Tier 3 lenses]
applied established concepts from training data.
## Headline finding
[One paragraph: the single most important thing the panel found,
supported by the convergence count and grounding tier]
## Where the panel converged (strong signal)
[Findings with 3+ framework convergence. Each finding gets:]
- **Finding**: [what]
- **Frameworks**: [which lenses converged]
- **Grounding**: [Tier N — source references]
- **Why it matters**: [impact if unaddressed]
- **Recommended action**: [specific, with file references]
## Where the panel converged (moderate signal)
[2-framework convergence findings, same format but briefer]
## Productive disagreements
[Named tensions between frameworks, with both positions and their
sources stated fairly. The user decides — the panel does not.]
## Prioritized action list
1. [Action] — [files] — [effort] — [frameworks + tiers]
2. ...
## What not to change
[Things multiple frameworks explicitly defended. Prevents the user
from "fixing" things that are already right.]
## Per-lens appendix
[The full review from each lens, with grounding tier and source
citations, for the user who wants the raw framework perspectives.]
Save to docs/reviews/YYYY-MM-DD-expert-panel-<topic>.md if docs/reviews/ exists, otherwise present inline.
Exit: Report delivered.
Use AskUserQuestion with:
Each lens defines: the analytical framework it applies, the published sources it draws from, the evaluative criteria it uses, the vocabulary that makes its findings distinct, and what it looks for that other lenses do not.
The lens names are the FRAMEWORK names. The authors are SOURCES, not personas.
learning-design — Learning Science & Cognitive LoadFramework: How people actually learn — attention management, cognitive load theory, progressive disclosure, transfer vs. retention, and practice installation.
Key sources:
Evaluative criteria:
Vocabulary: attention residue, metacognitive scaffolding, progressive disclosure, desirable difficulty, retrieval practice, transfer, concept before implementation.
evolutionary-design — Design Quality & Refactoring CraftFramework: Evolutionary software design — making change safe, growing systems incrementally, using tests as specification, and recognizing design smells as signals.
Key sources:
Evaluative criteria:
Vocabulary: evolutionary design, reversible decisions, refactoring toward, code smell, duplication of knowledge, rule of three, strangler fig, YAGNI, making change safe.
practitioner-reality — Monday Morning SurvivabilityFramework: Engineering-team-level reality check — the gap between conference advice and what actually works when you're shipping with a real team under real constraints.
Key sources:
Evaluative criteria:
Vocabulary: shipping reality, evidence-based, hype detection, adoption barriers, velocity pressure, "what actually works."
agent-engineering — Agent Reliability & Eval MethodologyFramework: How frontier labs and serious practitioners think about AI agent scaffolding — harness design, eval-driven development, approval loops, context management, and honest capability claims.
Key sources:
Evaluative criteria:
Vocabulary: harness, scaffolding, eval harness, approval loop, sandboxing, long-horizon drift, context budget, effective capability, verification boundary.
tool-craft — Concrete Examples & Tool AffordancesFramework: Practitioner-grade tool guidance — show the thing running, name the affordances, demonstrate failure, install a learning practice that outlasts any specific tool version.
Key sources:
Evaluative criteria:
Vocabulary: concrete examples, "show me it running," tool affordances, prompt transparency, failure cases, before/after pairs, learning practice, ecosystem velocity.
behavioral-change — Habit Installation & Operating ModelFramework: How organizations and teams actually change behavior — operating system design, autonomy with constraints, the difference between workshops that install habits and ones that entertain.
Key sources:
Evaluative criteria:
Vocabulary: operating model, behavioral installation, loose rules / tight feedback, self-organization, container design, distributed authority, constraint-based autonomy, environment design.
systems-dynamics — Feedback Loops & Leverage PointsFramework: Systems thinking applied to software delivery — stocks and flows, feedback loops, leverage points, and the conditions for continual learning.
Key sources:
Evaluative criteria:
Vocabulary: feedback loop, leverage point, stock and flow, delay, reinforcing loop, balancing loop, three ways, DORA metrics, batch size.
narrative-craft — Story Arc & Technical CommunicationFramework: How to communicate technical ideas so they land — narrative structure, audience contract, "show don't tell," and the difference between information transfer and understanding transfer.
Key sources:
Evaluative criteria:
Vocabulary: narrative arc, hook, "what is" / "what could be," show don't tell, audience contract, call to action, concrete analogy.
Before delivering the report:
/deep-thought:review instead./deep-thought:review./deep-thought:plan → /marvin:work./deep-thought:review for single-lens evaluation.../knowledge/critical-evaluation.md — Evidence-based evaluation, uncertainty flagging../knowledge/socratic-patterns.md — CoVe technique for verifying findings../knowledge/decision-frameworks.md — Tradeoff evaluation for the prioritized action list