From solo
This skill should be used when the user asks to 'test my idea', 'should I build', or 'validate idea'.
npx claudepluginhub jamon8888/cc-suite --plugin SoloThis skill uses the workspace's default tool permissions.
Answers two questions in one session:
Provides UI/UX resources: 50+ styles, color palettes, font pairings, guidelines, charts for web/mobile across React, Next.js, Vue, Svelte, Tailwind, React Native, Flutter. Aids planning, building, reviewing interfaces.
Fetches up-to-date documentation from Context7 for libraries and frameworks like React, Next.js, Prisma. Use for setup questions, API references, and code examples.
Analyzes multiple pages for keyword overlap, SEO cannibalization risks, and content duplication. Suggests differentiation, consolidation, and resolution strategies when reviewing similar content.
Share bugs, ideas, or general feedback.
Answers two questions in one session:
No spreadsheets, no scoring matrices. Just the minimum signal needed to decide.
Work through this checklist before committing any build time. Each question is Yes / No / Don't Know.
| Question | What a "Yes" looks like |
|---|---|
| Problem is real | You've heard at least 3 people describe it unprompted |
| Pain is acute | They're already doing something painful to solve it (workaround exists) |
| You can reach them | You know where they hang out (community, job board, Twitter, etc.) |
| They'd pay (or use it) | They've expressed frustration about not having a solution |
| You can build v1 alone | Core functionality is achievable in 2-4 weeks |
| Score | Verdict | Action |
|---|---|---|
| 4–5 Yes | GO | Pick a test (Part 2), confirm the riskiest assumption |
| 2–3 Yes | LEARN MORE | Run Part 2 with a cheap test before committing |
| 0–1 Yes | STOP | The problem or reach is unclear — return to discovery |
If you hit "Don't Know" on problem or pain, that's your signal: run a test before anything else.
The rule: use the cheapest test that tells the hardest truth.
Don't build a prototype to validate what a landing page can answer. Don't write a landing page to validate what a Reddit post can answer.
| Test | What it answers | Time | Cost | Best when |
|---|---|---|---|---|
| Smoke Test / Fake Door | "Will people click to sign up?" | 1–3 days | Free–€50 | You want demand signal before building anything |
| Concierge MVP | "Will people pay / use if I do it manually?" | 3–7 days | Time only | You can fake the product with manual work |
| Wizard of Oz | "Will they use the workflow?" | 3–7 days | Time only | The UX matters but code doesn't yet |
| Narrative / Video | "Does the story land?" | 1–2 days | Free | You need stakeholder or user buy-in on a concept |
| Spike / Feasibility | "Can I actually build this?" | 1–2 days | Dev time | Technical risk is the biggest unknown |
Answer these two questions:
If the risk is demand ("do enough people want this?") → Smoke Test or Concierge If the risk is behavior ("will they actually go through the flow?") → Wizard of Oz If the risk is comprehension ("will they get what this does?") → Narrative / Video If the risk is technical ("can this be built at all?") → Spike
For each test, define three things before starting:
Test: [which method]
Hypothesis: [If I do X, then Y% of people will Z]
Pass criteria: [the number that means GO — e.g., 10 signups, 3 paying users]
Timeline: [by when]
Save to data/1-Projets/[project]/validation-results.md
| Result | Action |
|---|---|
| Pass criteria met | GO → move to /solo:build design |
| Partial signal | ITERATE → tweak the hypothesis, run again or try a different angle |
| No signal | PIVOT → return to /solo:build discover with the new insight |
| Adjacent opportunity spotted | PIVOT → the test revealed a better problem |
Don't build a prototype to validate a demand hypothesis. A landing page is 10× faster.
Don't confuse "they liked it" with "they'd use it." People are polite. A signup or a payment is signal. Positive feedback in a conversation is not.
Don't test multiple hypotheses at once. One test, one assumption. If three things must be true, test the one you're most uncertain about first.
Don't skip the test because you're confident. That's confirmation bias.
SENTINEL_ROOT = "${CLAUDE_PLUGIN_ROOT}/../sentinel-v8"
sentinel_installed = file_exists(f"{SENTINEL_ROOT}/.claude-plugin/plugin.json")
Skip this section entirely if sentinel_installed is False. Idea-test works identically without it.
Activate after the Part 1 checklist produces a GO or LEARN MORE verdict. Do NOT activate on STOP verdicts — if the idea is dead, don't add analysis overhead.
Step 1 — failure-finder (pre-mortem)
Invoke the failure-finder agent with this specific frame:
"It's 12 months later. You built [idea]. Nobody is paying for it (or using it, if free). The product exists but has no traction. What happened?"
Generate 5 failure modes. Prioritize:
Step 2 — questioner (bias check)
Load bias IDs: [3, 7, 14, 22] from ../sentinel-v8/skills/decision-hygiene/references/bias-catalog.yaml
Generate 3 questions max. If the checklist reasoning is sound, say so instead.
Append to the standard GO/LEARN MORE verdict:
⚠️ Pre-mortem — before you start, stress-test these:
Failure mode 1 (HIGH): [what went wrong] → [early warning signal]
Failure mode 2 (MEDIUM): [what went wrong] → [early warning signal]
Answer these first:
Q1: [questioner question]
Q2: [questioner question]
If you can answer both honestly → proceed. If not → run the cheaper test first.
Standard GO / LEARN MORE / STOP verdict with anti-patterns. No change.