Use when applying strategic frameworks such as SWOT, RICE, OKR, or GTM
From second-claude-codenpx claudepluginhub unclejobs-ai/second-claude-code --plugin second-claude-codeThis skill uses the workspace's default tool permissions.
gotchas.mdreferences/challenge-round.mdreferences/frameworks/ansoff.mdreferences/frameworks/battlecard.mdreferences/frameworks/gtm.mdreferences/frameworks/journey-map.mdreferences/frameworks/lean-canvas.mdreferences/frameworks/north-star.mdreferences/frameworks/okr.mdreferences/frameworks/persona.mdreferences/frameworks/pestle.mdreferences/frameworks/porter.mdreferences/frameworks/prd.mdreferences/frameworks/pricing.mdreferences/frameworks/rice.mdreferences/frameworks/swot.mdreferences/frameworks/value-prop.mdDesigns and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
Never conclude first. Evidence determines the conclusion.
Apply a strategic framework to a topic, then stress-test it with a challenge pass.
| Framework | Best For |
|---|---|
swot | Situational assessment |
rice | Prioritization scoring |
okr | Goal-setting |
prd | Product requirements |
lean-canvas | Startup or product validation |
persona | User definition |
journey-map | Experience analysis |
pricing | Pricing strategy |
gtm | Go-to-market planning |
north-star | Metric alignment |
porter | Competitive forces |
pestle | Macro-environment scan |
ansoff | Growth direction |
battlecard | Competitive comparison |
value-prop | Value proposition design |
skills/analyze/references/frameworks/{framework}.md.strategist to apply the framework with evidence and clear implications, enforcing Source Requirements.devil-advocate to attack the weakest 3 points. At thorough depth with mmbridge detected, also dispatch mmbridge debate in parallel.thorough depth, run a second challenge pass focused on sourcing quality.| Flag | Values | Default |
|---|---|---|
--framework | listed above | auto-detect (see below) |
--with-research | flag | off |
--depth | quick|standard|thorough | standard |
--skip-challenge | flag | off |
--lang | ko|en | ko |
When --framework is not specified, detect from the prompt:
swot and state the assumption.When analyzing competitors, products, or external entities:
[unverified].| Level | Label | What happens | When to use |
|---|---|---|---|
| 1 | quick | Apply the template only. No research, no challenge round. | Internal brainstorming, time-boxed sessions |
| 2 | standard | Apply template with evidence-enforcement + one challenge round. | Default for most analyses |
| 3 | thorough | Full research pass (web search + source code reading) before applying template. Two challenge rounds + mmbridge debate (if available). Second round specifically attacks sourcing quality. | Competitive analysis, strategic decisions, publishable artifacts |
quick maps to --depth quickstandard maps to --depth standard (default)thorough maps to --depth thorough# {Framework} Analysis: {topic}
## Analysis
(framework sections with evidence)
## Challenge
- 3 weakest points
- blind spots and assumptions
## Balanced Insight
(synthesis)
## Recommended Actions
1. ...
2. ...
3. ...
.captures/analyze-{framework}-{slug}-{YYYY-MM-DD}.md{slug} = topic lowercased, spaces to hyphens, max 40 charsMandatory at standard (1 round) and thorough (2 rounds). See references/challenge-round.md for the full protocol.
When mmbridge is detected (see references/mmbridge-integration.md), analyze can dispatch mmbridge debate during the challenge round.
--depth standard: skip mmbridge debate--depth thorough: run it in parallel with the internal devil-advocateAt Step 4 (Challenge Round 1), also run via Bash:
mmbridge debate "<proposition from analysis>" --rounds 2 --json > /tmp/mmbridge-debate-${RUN_ID}.json
FOR and AGAINSTAGAINST arguments are merged into the challenge round as additional attack vectorsFOR arguments that contradict internal devil-advocate findings are noted as "disputed by external model"mmbridge debate costs more than a single internal challenge, so only activate it at thorough depth.
analyze spans both Plan and Do phases:
Add --context plan|do to override auto-detected PDCA phase.
analyst: { model: sonnet, constraint: "pre-process raw research data into structured patterns before framework application" }
strategist: { model: sonnet, constraint: "apply framework, cite evidence, fill required sections" }
devil-advocate: { model: sonnet, constraint: "attack weakest 3 points, surface blind spots" }