Code review with confidence-based filtering. Modes: quick (lint+types), standard (Opus), thorough (parallel agents). Auto-fixes ≥95% confidence issues.
Reviews uncommitted changes with confidence-based filtering and auto-fixes high-confidence issues.
/plugin marketplace add GGPrompts/TabzChrome/plugin install conductor@tabz-chromeReview uncommitted changes with confidence-based filtering. Only reports issues with ≥80% confidence. Auto-fixes issues with ≥95% confidence.
/conductor:code-review # Standard review (Opus)
/conductor:code-review --quick # Fast: lint + types + secrets only
/conductor:code-review --thorough # Deep: parallel specialized reviewers
/conductor:code-review <issue-id> # Review for specific issue
--quick)Fast checks only - use for trivial changes (docs, config, comments):
echo "=== Quick Code Review ==="
# Lint check
LINT_ERRORS=$(npm run lint 2>&1 | grep -c "error" || echo "0")
# Type check
TYPE_ERRORS=$(npx tsc --noEmit 2>&1 | grep -c "error" || echo "0")
# Secret scan
SECRETS=$(grep -rn "api[_-]?key\|secret\|password\|token" --include="*.ts" --include="*.tsx" src/ 2>/dev/null | grep -v "process.env\|import\|type\|interface" | wc -l)
if [ "$LINT_ERRORS" -eq 0 ] && [ "$TYPE_ERRORS" -eq 0 ] && [ "$SECRETS" -eq 0 ]; then
echo '{"passed": true, "mode": "quick", "summary": "Quick review passed"}'
else
echo "Lint errors: $LINT_ERRORS, Type errors: $TYPE_ERRORS, Potential secrets: $SECRETS"
echo '{"passed": false, "mode": "quick", "summary": "Quick checks failed"}'
fi
Spawns the conductor:code-reviewer agent (Opus) for thorough single-agent review:
Task(
subagent_type="conductor:code-reviewer",
prompt="Review uncommitted changes in $(pwd) for issue ${ISSUE_ID:-unknown}"
)
The agent will:
passed boolean--thorough)Spawns parallel specialized reviewers for comprehensive analysis. Use for:
## Thorough Review Pipeline
Spawn these agents in PARALLEL (single message with multiple Task calls):
### Agent 1: CLAUDE.md Compliance
Task(
subagent_type="general-purpose",
model="haiku",
prompt="Read all CLAUDE.md files in this repo. Then review the git diff for violations of any explicit rules. Return JSON: {violations: [{file, line, rule, confidence}]}"
)
### Agent 2: Bug Scanner
Task(
subagent_type="general-purpose",
model="sonnet",
prompt="Review git diff HEAD for bugs: null access, race conditions, logic errors, off-by-one. Focus ONLY on changed lines. Return JSON: {bugs: [{file, line, issue, confidence}]}"
)
### Agent 3: Silent Failure Hunter
Task(
subagent_type="general-purpose",
model="sonnet",
prompt="Review git diff HEAD for silent failures: empty catch blocks, swallowed errors, fallbacks that hide problems, missing error logging. Return JSON: {failures: [{file, line, issue, confidence}]}"
)
### Agent 4: Security Scanner
Task(
subagent_type="general-purpose",
model="haiku",
prompt="Review git diff HEAD for security issues: XSS, injection, exposed secrets, insecure patterns. Return JSON: {security: [{file, line, issue, severity, confidence}]}"
)
After all agents complete:
| Score | Meaning | Action |
|---|---|---|
| 0 | False positive, pre-existing issue | Skip |
| 25 | Might be real, can't verify, not in CLAUDE.md | Skip |
| 50 | Real but minor nitpick, low impact | Skip |
| 75 | Likely real, but not certain | Skip |
| 80-94 | Verified OR explicit CLAUDE.md violation | Flag |
| 95-100 | Certain - confirmed bug or clear rule | Auto-fix |
The reviewer will skip these even if they look like issues:
// eslint-disable)All modes return JSON:
{
"passed": true,
"mode": "standard",
"summary": "Reviewed 5 files. Auto-fixed 2 issues. No blockers. Tests recommended.",
"claude_md_checked": ["CLAUDE.md"],
"auto_fixed": [
{"file": "src/utils.ts", "line": 45, "issue": "Unused import", "confidence": 98}
],
"flagged": [
{"severity": "important", "file": "src/api.ts", "line": 23, "issue": "Missing error handling", "confidence": 85, "rule": "CLAUDE.md requires try-catch"}
],
"blockers": [],
"needs_tests": true,
"test_assessment": {
"recommendation": "recommended",
"rationale": "New utility function with multiple code paths",
"suggested_tests": [
{"type": "unit", "target": "formatDate()", "cases": ["valid date", "null", "invalid"]}
],
"priority": "medium",
"auto_writable": true
}
}
Every review includes a needs_tests assessment:
| Field | Description |
|---|---|
needs_tests | true/false - whether tests are warranted |
recommendation | required (blocks), recommended (flag), optional, skip |
rationale | Why tests are/aren't needed |
suggested_tests | Specific test cases to write |
priority | high/medium/low urgency |
auto_writable | Can tests be auto-generated (pure functions, clear I/O) |
Recommendation levels:
required - Risk area, bug fix, complex logic - blocks pipelinerecommended - New functions, moderate changes - flag onlyoptional - Simple changes - note for considerationskip - Docs, config, formatting - no tests neededpassed: false if ANY of:
recommendation: "required") but not present/conductor:verify-build - Run build before review/conductor:run-tests - Run tests before review/conductor:commit-changes - Run after review passes/conductor:worker-done - Full pipeline (includes this)| Situation | Mode |
|---|---|
| Docs, README, comments | --quick |
| Normal feature/fix | (default) |
| Security-sensitive code | --thorough |
| Large refactor (>500 lines) | --thorough |
| Auth, payments, data handling | --thorough |
| Config changes only | --quick |
| Adding tests | --quick |