Use when creating or developing ideas before writing code or implementation plans - refines rough ideas into fully-formed designs through collaborative questioning with Shannon quantitative validation, alternative exploration, and incremental validation. Don't use during clear mechanical processes
Transforms rough ideas into validated designs through systematic questioning and quantitative complexity analysis.
npx claudepluginhub krzemienski/shannon-frameworkThis skill inherits all available tools. When active, it can use any tool Claude has access to.
Help turn ideas into fully formed designs and specs through natural collaborative dialogue with Shannon's quantitative rigor.
Core principle: Understand deeply, explore alternatives, validate quantitatively, document systematically.
Check project context first:
Ask questions one at a time to refine the idea:
Focus on understanding:
Propose 2-3 different approaches with trade-offs:
Example:
I see three approaches:
**Option 1: Microservices** (Recommended)
- Complexity: 0.75/1.00 (COMPLEX)
- Duration: 2-3 weeks
- Pros: Scalability, isolation
- Cons: Higher initial complexity
- **Why recommended**: Aligns with growth plans
**Option 2: Monolith**
- Complexity: 0.45/1.00 (MODERATE)
- Duration: 1 week
- Pros: Simpler, faster to build
- Cons: Scaling limitations
**Option 3: Hybrid**
- Complexity: 0.55/1.00 (MODERATE-COMPLEX)
- Duration: 1.5-2 weeks
- Pros: Balance of both
- Cons: Some compromise on both sides
Which appeals to you?
Once you understand what you're building:
Break into sections of 200-300 words:
Be ready to go back and clarify if something doesn't make sense
Shannon quantitative validation:
## Design Validation
**Complexity Analysis** (Shannon 8D):
- Backend: 60%
- Data: 25%
- Frontend: 10%
- DevOps: 5%
**Overall**: 0.65/1.00 (COMPLEX)
**Estimated Effort**: 80-100 hours
**Risk Assessment**:
- HIGH: Database migrations
- MEDIUM: API integration
- LOW: UI updates
**MCP Requirements**:
- puppeteer (functional testing)
- sequential (complex analysis)
- context7 (framework docs if needed)
Write validated design to docs/plans/YYYY-MM-DD-<topic>-design.md:
# [Feature] Design Document
**Date**: YYYY-MM-DD
**Status**: Approved
**Complexity**: 0.65/1.00 (COMPLEX)
## Overview
[2-3 sentences]
## Architecture
[Design sections validated through brainstorming]
## Shannon Analysis
**8D Complexity**: 0.65/1.00
- Backend: 60%
- Data: 25%
- Frontend: 10%
- DevOps: 5%
**Estimated Effort**: 80-100 hours
**Risk Assessment**:
- HIGH: Database migrations
- MEDIUM: API integration
- LOW: UI updates
**Validation Strategy**:
- Tier 1: Flow validation
- Tier 2: Artifact validation
- Tier 3: Functional (NO MOCKS)
**MCP Requirements**:
- puppeteer
- sequential
- context7 (if using React/Next.js)
## Next Steps
- [ ] Create implementation plan (use writing-plans skill)
- [ ] Execute plan (use executing-plans or intelligent-do)
Commit the design document to git
Shannon tracking: Save to Serena:
serena.write_memory(f"designs/{design_id}", {
"feature": feature_name,
"complexity": 0.65,
"domain_distribution": {
"backend": 0.60,
"data": 0.25,
"frontend": 0.10,
"devops": 0.05
},
"estimated_hours": 90,
"risks": {"HIGH": 1, "MEDIUM": 1, "LOW": 1},
"approved": True,
"timestamp": ISO_timestamp
})
Ask: "Ready to set up for implementation?"
Two options:
Option 1: Write detailed plan (systematic approach)
Option 2: Direct execution (automatic approach)
During brainstorming, track:
brainstorm_metrics = {
"questions_asked": 12,
"alternatives_explored": 3,
"design_iterations": 2,
"sections_validated": 5,
"final_complexity": 0.65,
"estimated_hours": 90,
"confidence": 0.85, # How confident in the design
"timestamp": ISO_timestamp
}
serena.write_memory(f"brainstorm/{session_id}", brainstorm_metrics)
Learn from patterns:
# Query historical brainstorms
similar = serena.query_memory("brainstorm/*/complexity~0.6")
# Provide context
avg_duration = average([s["estimated_hours"] for s in similar])
# "Similar designs (complexity ~0.6) took average 85 hours"
This skill leads to:
This skill uses:
Shannon integration:
Brainstorming is design with Shannon's quantitative validation.
Not "this design looks good" - "this design: complexity 0.65, 90 hours, 3 HIGH risks mitigated, validated through 5 sections".
Systematic questioning + quantitative validation = reliable designs.
Activates when the user asks about AI prompts, needs prompt templates, wants to search for prompts, or mentions prompts.chat. Use for discovering, retrieving, and improving prompts.
Search, retrieve, and install Agent Skills from the prompts.chat registry using MCP tools. Use when the user asks to find skills, browse skill catalogs, install a skill for Claude, or extend Claude's capabilities with reusable AI agent components.
This skill should be used when the user asks to "create a hook", "add a PreToolUse/PostToolUse/Stop hook", "validate tool use", "implement prompt-based hooks", "use ${CLAUDE_PLUGIN_ROOT}", "set up event-driven automation", "block dangerous commands", or mentions hook events (PreToolUse, PostToolUse, Stop, SubagentStop, SessionStart, SessionEnd, UserPromptSubmit, PreCompact, Notification). Provides comprehensive guidance for creating and implementing Claude Code plugin hooks with focus on advanced prompt-based hooks API.