npx claudepluginhub fegome90-cmd/multi-review --plugin multi-reviewWant just this command?
Then install: npx claudepluginhub u/[userId]/[slug]
Thorough 4-agent code review - balanced review for features
--evidence-modeThorough Review (mr-thorough)
Run a balanced 4-agent code review optimized for feature development.
Variables
--evidence-mode: Enable evidence-based validation (runs ruff/mypy for Layer 3)
Instructions
This command runs a comprehensive review without the preset questionnaire.
Agent Configuration
Preset: thorough (4 agents, ~2 minutes)
| Agent | Purpose |
|---|---|
feature-dev:code-reviewer | General code review with confidence scoring |
pr-review-toolkit:pr-test-analyzer | Test coverage quality and completeness |
pr-review-toolkit:silent-failure-hunter | Error handling and silent failures |
pr-review-toolkit:code-simplifier | Refactoring and simplification suggestions |
Execution Workflow
Step 1: Build project context (3-Layer Defense - Layer 1)
python3 "${CLAUDE_PLUGIN_ROOT}/scripts/context_detector.py" --context-json
This extracts:
- Python config: mypy strictness, ruff rules, type checking level
- Shell config: Scripts with
set -euo pipefail(strict mode) - Test config: Test framework, coverage settings
- Git metadata: Changed files, branch info
Pass this context to each agent prompt to enable context-aware filtering.
Step 2: Launch agents in parallel
For EACH agent, launch a background agent:
Task(
subagent_type="<agent_type>",
description="Code review: <agent_name>",
prompt="<CALIBRATED_PROMPT_FROM_CACHE>
PROJECT CONTEXT:
- Python type checking: <mypy_strictness>
- Shell strict mode files: <strict_mode_files>
- Test framework: <test_framework>
Review the code changes in this workspace. Return findings with:
- Severity (critical/important/suggestion)
- Confidence score (0-100 based on criteria)
- File path and line numbers
- Brief description
- Suggested fix (if applicable)",
run_in_background=true
)
Store the returned task_id for each agent.
Step 3: Wait for all agents to complete
Use TaskOutput(task_id=..., block=true, timeout=300000) for each agent.
Step 4: Filter and aggregate results (3-Layer Defense - Layers 2 & 3)
python3 "${CLAUDE_PLUGIN_ROOT}/scripts/context_detector.py" --context-json > /tmp/project_context.json
Then aggregate findings:
python3 "${CLAUDE_PLUGIN_ROOT}/scripts/finding_aggregator.py" \
--context-json /tmp/project_context.json \
--findings-json /tmp/findings.json \
${EVIDENCE_MODE:+--evidence-mode} \
--output-format markdown
Confidence Scoring Criteria:
| Score | Confidence | Description |
|---|---|---|
| 0 | False positive | Pre-existing issue or doesn't stand up to scrutiny |
| 25 | Low | Might be real, but couldn't verify - likely nitpick |
| 50 | Medium | Verified real issue, but nitpick or rarely hit |
| 75 | High | Very likely real issue that will be hit in practice |
| 100 | Certain | Definitely real, frequently hit, directly confirmed |
Categorize findings:
- Critical Issues (filtered_confidence: 75-100)
- Important Issues (filtered_confidence: 50-74)
- Suggestions (filtered_confidence: 25-49)
- Suppressed (with reason logged)
Step 5: Present findings
Display categorized findings with agent attribution and confidence scores.
Step 6: Ask for next action
Use AskUserQuestion to offer:
Apply fixes- Runsuperpowers:receiving-code-reviewto process feedbackPlan fixes- Runsuperpowers:brainstormingto plan implementationDebug issues- Runsuperpowers:systematic-debuggingif critical problems foundStop- End command (manual fixes later)
Usage Examples
# Thorough review before PR
/mr-thorough
# With evidence validation
/mr-thorough --evidence-mode
When to Use
- Before creating a PR
- After implementing a new feature
- When code touches multiple areas
- Medium-complexity changes
See Also
/multi-review- Full interactive review with preset selection/mr-quick- 2-agent fast review/mr-comprehensive- 7-agent complete review