Invoked by /delivery command. Orchestrates feature delivery with Phase 0 classification, testing gates, and Superpowers/Ralph Loop pipeline selection. Do not invoke directly — use the /delivery command.
From hypershipnpx claudepluginhub hgrafa/hypership --plugin hypershipThis skill uses the workspace's default tool permissions.
acceptance-gate-prompt.mdbug-as-test-prompt.mdDesigns and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
Orchestrate feature delivery choosing the right strategy for the task. Your audience is senior engineers who make architectural decisions — be direct, technical, and skip introductory explanations.
Before evaluating delivery strategy, classify the prompt.
| Type | Criteria | Behavior |
|---|---|---|
feature | New functionality, no bug mention, no pure refactor | Proceed. Flag: acceptance_test_gate = true |
bugfix | Mentions error, bug, break, regression | Ask user to describe expected vs actual. Flag: bug_as_test_gate = true |
chore | Refactoring, dependency updates, config changes, performance tuning — no new behavior, no bug | Suggest /removedebt if pure refactoring. If user insists, proceed with standard TDD only — no extra gates. |
mixed | Contains feature + fix, or feature + chore | Decompose (see below) |
overloaded | 3+ distinct concerns, or scope too vague | Do not execute. Help refine. |
Decompose and present:
"I identified [N] distinct work items in this prompt:
- [type]: [description]
- [type]: [description]
Options:
- Sequential: [recommended order with reasoning]
- Parallel: independent deliveries via subagents (if no dependency)
Which do you prefer?"
After choice, each item enters the pipeline independently with its own type classification and corresponding gates.
"This prompt has [N] distinct concerns: [list]. For maximum quality, I propose breaking into separate deliveries. Which do you want to tackle first?"
If user insists on everything together, force decomposition into independent stories during brainstorm.
Phase 0 classification applies regardless of delivery mode:
Before starting, evaluate the task against these criteria:
TASK ARRIVES
│
▼
Can you define clear acceptance criteria
AND the task decomposes into >3 independent stories
AND the user wants AFK/autonomous execution?
│
YES ──▶ RALPH LOOP MODE
│ PRD-driven, autonomous, fresh context per iteration.
│ Best for: greenfield features, migrations, batch changes.
│
NO
│
▼
SUPERPOWERS MODE (default)
brainstorm → plan → subagent-driven-dev → finish
Best for: everything else. Complex logic, integrations,
features needing human judgment mid-implementation.
When in doubt, use Superpowers. Ralph is a power tool for specific scenarios; Superpowers is the general-purpose pipeline.
Ask the user which mode they prefer if the task fits both. Present the tradeoff concisely:
"This could run as a Ralph loop (autonomous, fresh context per story, good for AFK) or Superpowers pipeline (interactive, two-stage review per task, higher precision). Which do you prefer?"
Before starting, detect available tools silently:
Superpowers (required):
ls ~/.claude/plugins/cache/ | grep -i superpowersContext7 (optional):
claude mcp list 2>/dev/null | grep -i context7docs-researcher agent for API lookups
during implementation instead of relying on training knowledge.Memory (optional, any of these):
claude-mem plugin → use its MCP search toolsmemsearch plugin → use its query tools~/.claude/projects/*/memory/GitHub/GitLab MCP (optional):
claude mcp list 2>/dev/null | grep -i githubRalph Loop (optional):
scripts/ralph/ralph.shReport to user concisely:
"Env: Superpowers ✅ | Context7 [✅|❌] | GitHub [✅|❌] | Memory [✅|❌] | Ralph [✅|❌]"
Then invoke superpowers:using-superpowers to activate the skill system.
Follow the standard Superpowers pipeline. The skills activate automatically once brainstorming completes:
superpowers:brainstorming → design doc savedsuperpowers:using-git-worktrees → isolated branchsuperpowers:writing-plans → granular tasks in docs/plans/superpowers:subagent-driven-development → implement + two-stage reviewsuperpowers:finishing-a-development-branch → merge/PR/keep/discardGates activate based on Phase 0 classification flags. They modify
how the Superpowers subagent-driven-development skill dispatches
and reviews each task.
When bug_as_test_gate = true (bugfix):
./bug-as-test-prompt.md and append to the implementer promptBLOCKED_NON_REPRODUCIBLE
with 3 alternative approaches. Present to user for choice.BLOCKED_INFRA.
Spec reviewer validates the justification.NO BUGFIX WITHOUT EVIDENCE.When acceptance_test_gate = true (feature):
./acceptance-gate-prompt.md, replace the
[ACCEPTANCE_CRITERIA] placeholder with the actual criteria list from
brainstorm, and append to the spec reviewer promptWhen neither flag is set (chore):
After Superpowers finishes:
git tag delivery/YYYY-MM-DD-feature-namedocs/delivery-log.md:
## [date] feature-name
- Plan: docs/plans/plan-file.md
- Branch: branch-name
- Stories completed: N/N
- Status: merged | pending-review
removedebt entry in docs/debt-log.md. If > 5, suggest:
"You have N features since last debt removal. Consider running
/removedebt since last consolidation"
Check Ralph availability:
ls ~/.claude/plugins/cache/ | grep -i ralphscripts/ralph/ralph.sh existsIf neither exists, offer to set up:
/plugin install ralph-loop
Or manual setup with Superpowers brainstorming to generate the PRD.
Use superpowers:brainstorming to generate the PRD. This is where the
two tools combine: Superpowers' structured brainstorm produces a better
PRD than writing one manually.
After brainstorm, format as PRD:
{
"branchName": "feat/feature-name",
"userStories": [
{
"id": "STORY-1",
"title": "story title",
"acceptanceCriteria": ["criterion 1", "criterion 2"],
"passes": false
}
]
}
Rule: Each story must fit in a single context window. If it's too big, split it. Ask the user if unsure about granularity.
Present to user:
"Ralph loop configured with N stories. Options:
- HITL (human-in-the-loop): you watch each iteration, review, re-run
- AFK: autonomous with max-iterations safety. I notify on completion.
For first time with this feature scope, I recommend HITL."
Ralph handles the loop. After completion:
Regardless of mode, before starting:
Present to user after delivery completes:
✅ Feature implemented and tested
✅ Plan saved to docs/plans/
✅ Delivery logged in docs/delivery-log.md
✅ Branch merged / PR created
⚠️ Features since last /removedebt: [N]
[If > 5: "Recommend running /removedebt"]