From strategy-consultant
Plan, guide, and process expert interviews to confirm and enhance research findings. Use when someone asks to "prepare for expert interviews", "create an interview guide", "who should we interview", "process interview notes", "extract insights from interviews", or when the analytical workflow includes expert interviews after the research phase. Also trigger when someone uploads expert interview transcripts or notes that need to be analyzed and integrated into the research base.
npx claudepluginhub chipalexandru/strategy-consultantThis skill uses the workspace's default tool permissions.
Provides UI/UX resources: 50+ styles, color palettes, font pairings, guidelines, charts for web/mobile across React, Next.js, Vue, Svelte, Tailwind, React Native, Flutter. Aids planning, building, reviewing interfaces.
Fetches up-to-date documentation from Context7 for libraries and frameworks like React, Next.js, Prisma. Use for setup questions, API references, and code examples.
Builds 3-5 year financial models for startups with cohort revenue projections, cost structures, cash flow, headcount plans, burn rate, runway, and scenario analysis.
Expert interviews come AFTER the public research phase. Their purpose is to confirm findings, fill information gaps, add precision to estimates, and surface insights that public data cannot provide. Expert interview data carries high weight (CS-2) because it comes from practitioners with direct domain knowledge.
This skill covers three phases: planning who to interview, creating interview guides informed by research findings, and processing interview outputs back into the research flow.
This skill activates in two scenarios:
If the user indicated during the Data Source Inquiry (Phase 2.5) that expert interviews would be available, this skill should be invoked after the research phase completes.
Based on the Precision Anchor, the validated research findings, and the information gaps identified by the research-validator, recommend specific expert profiles.
Review the validator's output — specifically:
For each gap, write a one-sentence description of what the expert needs to provide.
For each priority gap, recommend a specific expert profile. Be precise — not "someone in the industry" but a targeted description:
Format for each expert profile:
Expert Profile #[N]
Role: [e.g., "Former VP of Supply Chain at a top-3 European FMCG company"]
Why this expert: [Which specific information gap they are best positioned to close]
Key qualification: [What direct experience makes them credible — e.g., "managed distribution across 5+ EU markets in the last 3 years"]
Priority: [HIGH / MEDIUM] — HIGH if the gap blocks a critical claim, MEDIUM if it would strengthen an already-supported finding
Company type: [e.g., "competitor", "customer", "supplier", "regulator", "industry body"]
Rank the expert profiles by impact on the engagement. Recommend 3-5 experts maximum, prioritized by:
Present the expert profiles to the user and ask for confirmation before creating interview guides.
Create concise, problem-specific interview guides — NOT generic questionnaires. Each guide is tailored to the Precision Anchor, the specific gaps from the research phase, and the expert's profile.
Before writing questions, identify:
Guide structure:
INTERVIEW GUIDE — [Expert Profile Description]
Engagement: [Problem statement — one line]
Date: [To be filled]
Interviewer: [To be filled]
CONTEXT FOR THE INTERVIEWER
[2-3 sentences summarizing what the research has found so far and what this interview aims to confirm, challenge, or deepen. This helps the interviewer understand the strategic context without biasing the expert.]
OPENING (2-3 minutes)
- Brief introduction of the engagement context (without leading the expert)
- Ask the expert to describe their relevant experience in their own words
- Establish the time horizon and scope they are most knowledgeable about
KEY QUESTIONS (20-30 minutes)
[5-8 questions, each explicitly mapped to an information gap]
Q1: [Question]
→ Gap this addresses: [Which information gap from the research phase]
→ What we already know: [Brief summary of public research on this point]
→ What we need from the expert: [Specific data point, validation, or insight]
→ Follow-up probes: [2-3 follow-up questions if the initial answer is vague or incomplete]
Q2: [Question]
→ Gap this addresses: ...
...
VALIDATION SECTION (5-10 minutes)
[Present 2-3 specific findings from the public research and ask the expert to confirm, challenge, or nuance them. Frame as: "Our research suggests [X]. Does that align with your experience? What would you adjust?"]
CLOSING (2-3 minutes)
- "What haven't I asked about that you think is important for this question?"
- "Who else would you recommend we speak with?"
- "Is there any data or documentation you could share that would help?"
Create a traceability table showing which interview question addresses which information gap:
| Question | Information Gap | Current Best Evidence | CS Score of Current Evidence | What Expert Input Would Provide |
|---|
This ensures no critical gap is missed and helps prioritize if the interview runs short.
When the user uploads interview notes or transcripts, process them systematically to extract findings and integrate them into the research base.
Read the interview notes in full. For every substantive claim the expert makes, extract it as a discrete data point:
[1] Claim: "[Exact claim in the expert's words]"
Expert: [Name, title, company]
Context: [What prompted this claim — the question asked or topic discussed]
Type: [FACT — based on the expert's direct experience / OPINION — the expert's judgment or prediction / HEARSAY — something the expert heard but did not directly experience]
Separate facts from opinions rigorously. An expert saying "we saw 15% cost reduction when we switched suppliers" is a FACT (their direct experience). An expert saying "I think the market will grow 20%" is an OPINION. An expert saying "I heard Company X is planning to enter" is HEARSAY.
Expert interview claims are scored as follows:
This is the critical step. For every expert claim, compare it against the public research findings:
Agreement: Expert confirms a public finding → Upgrade confidence. Note: "Corroborated by expert interview with [name], [company]." If the public finding was CS-3, it may now be upgradeable with CS-2 expert corroboration.
Enhancement: Expert adds precision or detail to a public finding → Record the enhanced version. Example: public research says "market is growing"; expert says "market grew 18% in our segment last year, driven by regulatory changes." The expert version is more precise and actionable.
Conflict: Expert contradicts a public finding → This requires careful adjudication:
Net new insight: Expert provides information that public research did not uncover → Add as a new finding with CS-2 (or CS-1 if verifiable company data). Note that this is "expert-sourced, not publicly corroborated."
## Interview Summary
Expert: [Name, title, company, date of interview]
Relevance to engagement: [One sentence on why this expert was selected]
## Key Findings from Interview
[Numbered list of the most important claims, each with type (FACT/OPINION/HEARSAY) and CS score]
## Cross-Reference Against Public Research
### Confirmed Findings
| # | Public Finding | Expert Confirmation | Effect on Confidence |
|---|---------------|--------------------|--------------------|
### Enhanced Findings
| # | Original Public Finding | Enhanced Version (from expert) | Improvement |
|---|------------------------|-------------------------------|-------------|
### Conflicts
| # | Public Finding (CS score) | Expert Claim (CS score) | Recommended Resolution | Reasoning |
|---|--------------------------|------------------------|----------------------|-----------|
### Net New Insights
| # | Expert Claim | CS Score | Gap It Fills | Publicly Corroborated? |
|---|-------------|----------|-------------|----------------------|
## Updated Source Registry Entries
[New or modified Source Registry entries resulting from this interview, in the standard format for inclusion in Research Notes]
[N] Data point: "[claim]"
Source: Expert interview — [Name], [Title], [Company], [Date]
URL: [N/A — expert interview]
CS Score: [CS-1 / CS-2 / CS-3]
Verbatim from source: "[Exact quote from interview notes]"
## Impact on Information Gaps
| Original Gap | Status After Interview | Notes |
|-------------|----------------------|-------|
[CLOSED / PARTIALLY CLOSED / STILL OPEN]
## Recommendations
[Any follow-up research needed based on new information or unresolved conflicts. Any additional experts to interview.]
After processing, the expert interview output should be: