This skill should be used when the user wants to review a document from different stakeholder perspectives, simulate how different audiences would read a document, or get multi-perspective feedback before distribution. Common triggers include "review document as stakeholder", "stakeholder review", "reader review", "read as executive", "review from technical perspective", "does this document work for [audience]", "get feedback on this document", "what would [role] think of this", "check if this is ready for stakeholders", or "simulate different readers".
From cogni-copywritingnpx claudepluginhub cogni-work/insight-wave --plugin cogni-copywritingThis skill is limited to using the following tools:
references/personas/cdo-utility.mdreferences/personas/cmo-provider.mdreferences/personas/end-user.mdreferences/personas/executive.mdreferences/personas/legal.mdreferences/personas/marketing.mdreferences/personas/technical.mdreferences/synthesis-protocol.mdGuides browser automation with Playwright, Puppeteer, Selenium for e2e testing and scraping. Teaches reliable selectors, auto-waits, isolation to fix flaky tests.
Provides checklists to review code for functionality, quality, security, performance, tests, and maintainability. Use for PRs, audits, team standards, and developer training.
Enforces A/B test setup with gates for hypothesis locking, metrics definition, sample size calculation, assumptions checks, and execution readiness before implementation.
MANDATORY: ALL German-specific characters MUST be preserved exactly as written. NEVER convert to ASCII equivalents (ae, oe, ue, ss). This applies to ALL text: body, headers, citations, technical terms.
MANDATORY: ALL citation markers and their URLs MUST be preserved exactly as written. Removing or omitting citations is a CRITICAL FAILURE. Citations are evidence markers for audit trail integrity.
DO NOT modify diagram placeholders, figure references, figure captions, Obsidian embeds, or kanban tables. Preserve exactly as-is.
Initialize TodoWrite with these 6 steps, then execute sequentially:
Extract from user request or agent invocation:
FILE_PATH: Absolute path to markdown document (required)PERSONAS: Array of perspectives to simulate (default: all five generic)
executive, technical, legal, marketing, end-user.md file in references/personas/ is a valid persona name (e.g. cdo-utility, cmo-provider)AUTO_IMPROVE: Whether to apply improvements directly (default: true)Validate:
Load persona references:
FOR EACH persona IN PERSONAS:
READ: references/personas/{persona}.md
Before any analysis, create a backup:
dir=$(dirname "${FILE_PATH}")
filename=$(basename "${FILE_PATH}")
backup_path="${dir}/.${filename}.pre-reader-review"
cp "${FILE_PATH}" "${backup_path}"
Report: Backup created: {backup_path}
Launch one Task agent per persona to analyze the document in parallel. Each persona agent reads the document and produces structured feedback from their perspective.
For each persona, launch a Task agent with this prompt:
You are a {PERSONA_NAME} stakeholder reading a document. Your job is to evaluate it from your specific perspective and produce structured feedback.
DOCUMENT PATH: {FILE_PATH}
Read the document, then evaluate using the criteria from your persona profile.
PERSONA PROFILE:
{content from references/personas/{persona}.md}
INSTRUCTIONS:
1. Read the entire document carefully
2. Evaluate against each criterion in your profile
3. Generate 3-5 questions a real {PERSONA_NAME} stakeholder would ask after reading
4. Identify specific concerns with line references where possible
5. Provide concrete improvement recommendations
OUTPUT FORMAT (JSON only):
{
"perspective": "{persona}",
"score": <0-100>,
"questions": ["Question 1?", "Question 2?", ...],
"concerns": ["Concern with specific reference", ...],
"recommendations": [
{"priority": "CRITICAL|HIGH|OPTIONAL", "action": "Specific improvement to make", "location": "Section or paragraph reference"}
],
"strengths": ["What works well", ...]
}
Agent configuration per persona:
After all persona agents return, synthesize their feedback.
Load synthesis protocol:
READ: references/synthesis-protocol.md
Synthesis process:
Synthesis output:
{
"persona_scores": {"executive": 78, "technical": 85, ...},
"overall_score": 82,
"cross_cutting_themes": [
{"theme": "Missing quantification", "personas": ["executive", "marketing"], "priority": "CRITICAL"}
],
"merged_recommendations": [
{"priority": "CRITICAL", "action": "...", "sources": ["executive", "marketing"]},
{"priority": "HIGH", "action": "...", "sources": ["technical"]}
],
"all_questions": [
{"persona": "executive", "question": "What's the expected ROI timeline?"},
...
]
}
If AUTO_IMPROVE is true (default):
Apply ONE improvement pass based on synthesized recommendations.
Process:
Apply all CRITICAL recommendations:
Apply HIGH recommendations where feasible:
Log OPTIONAL recommendations without applying
Validate final document:
If validation fails: Revert to backup and report failure reason.
Present a comprehensive report to the user or calling agent.
User-facing format:
## Reader Review: {filename}
**Personas consulted:** {list}
**Overall score:** {score}/100
**Backup:** {backup_path}
**Improvements applied:** {count}
### Persona Scores
| Persona | Score | Top Concern |
|---------|-------|-------------|
| Executive | 78 | Missing ROI timeline |
| Technical | 85 | Vague implementation details |
| ... | ... | ... |
### Questions Your Stakeholders Would Ask
**Executive:**
1. What's the expected payback period?
2. What happens if we delay this decision?
**Technical:**
1. What are the system dependencies?
2. How does this handle failure scenarios?
...
### Improvements Applied
1. **CRITICAL:** Added decision timeline (March 15, 2025) - raised Executive score to 88
2. **HIGH:** Broke dense paragraph into bullets - raised End-user score to 95
3. **OPTIONAL (logged):** Add vendor comparison table - requires external research
### Summary
{1-2 sentence summary of document readiness after improvements}
JSON format (for agent/skill callers):
{
"success": true,
"file": "{filename}",
"backup_path": "{backup_path}",
"personas_consulted": ["executive", "technical", ...],
"overall_score": 84,
"persona_results": [...],
"improvements_applied": 4,
"improvements_skipped": 1,
"questions": [...],
"protected_content_preserved": true
}
Each persona file defines:
Available personas (any .md file in this directory is valid):