From claudekit
Reviews architecture of written plans: scores data flow, failure modes, edge cases, test matrix, rollback safety (0-10 each) with citations; produces ranked fixes.
npx claudepluginhub duthaho/claudekit --plugin claudekitThis skill uses the workspace's default tool permissions.
The architecture-dimension reviewer for `plan-review`. Reads a plan and scores
Orchestrates parallel architecture and experience reviews of implementation plans, scores findings across dimensions like data flow and UX, consolidates ranked fixes for user approval and auto-application. Use after planning, before non-trivial coding.
Delivers systematic architecture reviews across 7 dimensions with scored reports, highlighting risks and recommendations.
Reviews implementation plans through quality lenses (architecture, security, test coverage, code quality, standards, usability) and collaboratively iterates before implementation.
Share bugs, ideas, or general feedback.
The architecture-dimension reviewer for plan-review. Reads a plan and scores
five concrete sub-dimensions on 0-10: data flow, failure modes, edge cases, test
matrix, and rollback safety. Every score must be paired with a finding citing the
plan task number or section that caused the score. The skill produces a ranked
fix list aligned with plan-review's consolidation step. Used by plan-review's
orchestrator, but invocable directly when only an architectural review is needed.
plan-review as one of its two parallel reviewersGoal: Build context before scoring.
Inputs: The plan file. Optionally: the spec it's derived from, the relevant codebase area.
Actions:
map-codebase mentally on the affected area: which files, which entry
points, which downstream services or queues.Output: A short pre-read note: Plan touches <areas>; primary risks I'll watch for: <list>.
Goal: Produce 5 scores with cited findings.
Inputs: The plan file plus pre-read notes.
Actions: For each sub-dimension below, score 0-10 and write at least one finding. Findings must cite the plan task number or section.
Data flow (0-10)
Failure modes (0-10)
Edge cases (0-10)
Test matrix (0-10)
Rollback safety (0-10)
NOT POSSIBLE and
gated behind a feature flag, dual-write, or backfill?Goal: Convert each finding into a concrete fix proposal.
Inputs: The findings from Step 2.
Actions:
<task or section> — change <X> to <Y> or Add <Z> to <task or section>.Output: A ranked list of fixes with cited targets in the plan.
Goal: Hand plan-review a clean, paste-ready report.
Inputs: Scores and ranked fixes.
Actions:
Produce a Markdown block with this structure:
## Architecture review
- Data flow: X/10 — <one-line justification>
- Failure modes: X/10 — <one-line justification>
- Edge cases: X/10 — <one-line justification>
- Test matrix: X/10 — <one-line justification>
- Rollback safety: X/10 — <one-line justification>
### Findings
- [Blocker] <finding>; fix: <fix>; cite: <task #>
- [Important] <finding>; fix: <fix>; cite: <task #>
- [Nice-to-have] <finding>; fix: <fix>; cite: <task #>
Hand back to plan-review for consolidation with the experience reviewer.
Output: The Markdown block.
| Excuse | Why it sounds reasonable | Why it's wrong | What to do instead | |||||||||||||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| "I'll score by gut feel — calibration is a waste of time." | Experienced reviewers do have calibrated guts. | Gut-feel scoring without rubric anchors produces "everything's a 7" output the user cannot act on. The rubric anchors exist so the score communicates which gap is open, not just that something feels off. | Use the 0/5/10 anchors above. If a sub-dimension feels like a 7, that's actually a "5 with one gap closed" — name the open gap, score 6 or 7, and write the finding for it. | |||||||||||||||||||||||
| "I'll skip the citations — the user can find the relevant tasks." | Plans are short; finding the cited task is fast. | Findings without citations leave the user to do the matching, and they will skip findings that take work to verify. The citation is the cheapest part of the review for the reviewer and the most expensive part to reconstruct for the consumer. | Cite the task number or plan section in every finding. Task 4 — failure mode for the cache miss is undefined not Cache failure modes are missing. | |||||||||||||||||||||||
| "Rollback for this is obviously the deploy team's problem." | Some rollbacks are operational, owned by SRE. | "Obviously theirs" is the line you say when you don't know what the rollback is. The author of the change knows what would need to be undone; the deploy team knows how to undo it. The plan needs the what, not the how. | Even if SRE owns execution, the plan author writes one line: "Rollback: revert ; re-run migration down; truncate ." If you can't write that line, escalate during review, don't skip during review.
Evidence Requirements
Red Flags
References
|