From sdlc
Execute plans with fresh subagents per task and two-stage review (spec compliance then code quality). Use over pair-build when tasks are independent, don't share files, and benefit from isolated context. Trigger: 'use subagents', 'dispatch workers', 'build with isolated agents'.
npx claudepluginhub jerrod/agent-plugins --plugin sdlcThis skill uses the workspace's default tool permissions.
Searches, retrieves, and installs Agent Skills from prompts.chat registry using MCP tools like search_skills and get_skill. Activates for finding skills, browsing catalogs, or extending Claude.
Searches prompts.chat for AI prompt templates by keyword or category, retrieves by ID with variable handling, and improves prompts via AI. Use for discovering or enhancing prompts.
Transforms Claude Code into autonomous agent system with persistent memory, scheduled operations, computer use, and task queuing. Use for continuous operation, scheduled tasks, or self-directing agent loops.
Log skill invocation:
AUDIT_SCRIPT=$(find . -name "audit-trail.sh" -path "*/sdlc/*" 2>/dev/null | head -1)
[ -z "$AUDIT_SCRIPT" ] && AUDIT_SCRIPT=$(find "$HOME/.claude" -name "audit-trail.sh" -path "*/sdlc/*" 2>/dev/null | sort -V | tail -1)
bash "$AUDIT_SCRIPT" log build sdlc:subagent-build started --context "$ARGUMENTS"bash "$AUDIT_SCRIPT" log build sdlc:subagent-build completed --context="<summary>"Execute plan by dispatching fresh subagent per task, with two-stage review after each: spec compliance review first, then code quality review.
Why subagents: You delegate tasks to specialized agents with isolated context. By precisely crafting their instructions and context, you ensure they stay focused and succeed at their task. They should never inherit your session's context or history — you construct exactly what they need. This also preserves your own context for coordination work.
Core principle: Fresh subagent per task + two-stage review (spec then quality) = high quality, fast iteration
NEVER truncate command output with | head, | tail, or | grep. Redirect to a tmp file (> /tmp/output.out 2>&1) and Read the file. One run, full output. Include this rule in every subagent prompt you construct.
digraph when_to_use {
"Have implementation plan?" [shape=diamond];
"Tasks mostly independent?" [shape=diamond];
"Stay in this session?" [shape=diamond];
"subagent-build" [shape=box];
"sdlc:pair-build" [shape=box];
"Manual execution or brainstorm first" [shape=box];
"Have implementation plan?" -> "Tasks mostly independent?" [label="yes"];
"Have implementation plan?" -> "Manual execution or brainstorm first" [label="no"];
"Tasks mostly independent?" -> "Stay in this session?" [label="yes"];
"Tasks mostly independent?" -> "Manual execution or brainstorm first" [label="no - tightly coupled"];
"Stay in this session?" -> "subagent-build" [label="yes"];
"Stay in this session?" -> "sdlc:pair-build" [label="no - parallel session"];
}
vs. sdlc:pair-build (parallel session):
digraph process {
rankdir=TB;
subgraph cluster_per_task {
label="Per Task";
"Dispatch implementer subagent (./implementer-prompt.md)" [shape=box];
"Implementer subagent asks questions?" [shape=diamond];
"Answer questions, provide context" [shape=box];
"Implementer subagent implements, tests, commits, self-reviews" [shape=box];
"Dispatch spec reviewer subagent (sdlc:spec-compliance-reviewer)" [shape=box];
"Spec reviewer subagent confirms code matches spec?" [shape=diamond];
"Implementer subagent fixes spec gaps" [shape=box];
"Dispatch code quality reviewer subagent (./code-quality-reviewer-prompt.md)" [shape=box];
"Code quality reviewer subagent approves?" [shape=diamond];
"Implementer subagent fixes quality issues" [shape=box];
"Mark task complete in TodoWrite" [shape=box];
}
"Read plan, extract all tasks with full text, note context, create TodoWrite" [shape=box];
"More tasks remain?" [shape=diamond];
"Dispatch final code reviewer subagent for entire implementation" [shape=box];
"Use sdlc:finish-branch" [shape=box style=filled fillcolor=lightgreen];
"Read plan, extract all tasks with full text, note context, create TodoWrite" -> "Dispatch implementer subagent (./implementer-prompt.md)";
"Dispatch implementer subagent (./implementer-prompt.md)" -> "Implementer subagent asks questions?";
"Implementer subagent asks questions?" -> "Answer questions, provide context" [label="yes"];
"Answer questions, provide context" -> "Dispatch implementer subagent (./implementer-prompt.md)";
"Implementer subagent asks questions?" -> "Implementer subagent implements, tests, commits, self-reviews" [label="no"];
"Implementer subagent implements, tests, commits, self-reviews" -> "Dispatch spec reviewer subagent (sdlc:spec-compliance-reviewer)";
"Dispatch spec reviewer subagent (sdlc:spec-compliance-reviewer)" -> "Spec reviewer subagent confirms code matches spec?";
"Spec reviewer subagent confirms code matches spec?" -> "Implementer subagent fixes spec gaps" [label="no"];
"Implementer subagent fixes spec gaps" -> "Dispatch spec reviewer subagent (sdlc:spec-compliance-reviewer)" [label="re-review"];
"Spec reviewer subagent confirms code matches spec?" -> "Dispatch code quality reviewer subagent (./code-quality-reviewer-prompt.md)" [label="yes"];
"Dispatch code quality reviewer subagent (./code-quality-reviewer-prompt.md)" -> "Code quality reviewer subagent approves?";
"Code quality reviewer subagent approves?" -> "Implementer subagent fixes quality issues" [label="no"];
"Implementer subagent fixes quality issues" -> "Dispatch code quality reviewer subagent (./code-quality-reviewer-prompt.md)" [label="re-review"];
"Code quality reviewer subagent approves?" -> "Mark task complete in TodoWrite" [label="yes"];
"Mark task complete in TodoWrite" -> "More tasks remain?";
"More tasks remain?" -> "Dispatch implementer subagent (./implementer-prompt.md)" [label="yes"];
"More tasks remain?" -> "Dispatch final code reviewer subagent for entire implementation" [label="no"];
"Dispatch final code reviewer subagent for entire implementation" -> "Use sdlc:finish-branch";
}
Note: When Test Strategy: pre-implementation is active, the "Per Task" cluster expands to three phases (interface design → test writing → implementation) before reaching the review stages. The standard flow (single implementer) is unchanged for tasks without this flag.
Use the least powerful model that can handle each role to conserve cost and increase speed.
Mechanical implementation tasks (isolated functions, clear specs, 1-2 files): use a fast, cheap model. Most implementation tasks are mechanical when the plan is well-specified.
Integration and judgment tasks (multi-file coordination, pattern matching, debugging): use a standard model.
Architecture, design, and review tasks: use the most capable available model.
Test writing (interface-to-test translation, constrained scope): use a cheap model. The test writer works from a narrow interface spec and doesn't need architectural judgment.
Task complexity signals:
Implementer subagents report one of four statuses. Handle each appropriately:
DONE: Proceed to spec compliance review.
DONE_WITH_CONCERNS: The implementer completed the work but flagged doubts. Read the concerns before proceeding. If the concerns are about correctness or scope, address them before review. If they're observations (e.g., "this file is getting large"), note them and proceed to review.
NEEDS_CONTEXT: The implementer needs information that wasn't provided. Provide the missing context and re-dispatch.
BLOCKED: The implementer cannot complete the task. Assess the blocker:
Never ignore an escalation or force the same model to retry without changes. If the implementer said it's stuck, something needs to change.
DONE_WITH_CONCERNS with CONFLICT_REASON: When using the pre-implementation test pipeline (below), the implementer may report a CONFLICT_REASON:
SPEC_AMBIGUITY → escalate to the developer for clarification. Pause the task.INTERFACE_MISMATCH → re-dispatch the test writer with the corrected interface spec from the implementer.When a plan task contains **Test Strategy:** pre-implementation, use a three-phase pipeline instead of the standard single-implementer flow:
Dispatch the implementer with modified instructions:
"Define the public API for this task. Write function/method signatures with types and error contracts. Do NOT implement any logic — use placeholder bodies that raise/throw. Commit this as the interface spec."
The implementer commits ONLY interface artifacts — no implementation bodies. The commit message should be: chore: define interface for task N
Dispatch a fresh test writer subagent using ./test-writer-prompt.md. Provide:
The test writer commits failing tests: test: write failing tests for task N
Use haiku model by default. Escalate to sonnet if the test writer reports NEEDS_CONTEXT or BLOCKED.
Re-dispatch the implementer with modified instructions:
"You will find failing tests already committed. Your job is to make them pass. You may ADD new tests but you may NOT modify, delete, or weaken any existing test assertions. If you believe a test is wrong, report DONE_WITH_CONCERNS with a CONFLICT_REASON — do not modify the test."
The implementer receives: plan task text, interface spec, and the committed test files.
Tasks without **Test Strategy:** pre-implementation use the existing single-implementer flow unchanged. Integration/wiring tasks, config changes, and tasks without clear public interfaces should NOT use this pipeline.
./implementer-prompt.md - Dispatch implementer subagent./test-writer-prompt.md - Dispatch pre-implementation test writer subagent./spec-reviewer-prompt.md - Dispatch spec compliance reviewer subagent./code-quality-reviewer-prompt.md - Dispatch code quality reviewer subagentNever:
If subagent asks questions:
If reviewer finds issues:
If subagent fails task:
Required workflow skills:
Subagents should use:
Alternative workflow: