From compound-engineering
[BETA] Stress-test an existing implementation plan and selectively strengthen weak sections with targeted research. Use when a plan needs more confidence around decisions, sequencing, system-wide impact, risks, or verification. Best for Standard or Deep plans, or high-risk topics such as auth, payments, migrations, external APIs, and security. For structural or clarity improvements, prefer document-review instead.
npx claudepluginhub apollostreetcompany/codex-compound --plugin compound-engineeringThis skill uses the workspace's default tool permissions.
**Note: The current year is 2026.** Use this when searching for recent documentation and best practices.
Generates structured implementation plans from feature descriptions or requirements, grounded in repo patterns and research. Deepens existing plans via interactive sub-agent review.
Creates structured plans for multi-step tasks including software features, implementations, research, or projects. Deepens plans via interactive sub-agent reviews.
Creates structured plans for multi-step tasks like software features, research workflows, events, or study plans. Deepens existing plans with interactive sub-agent reviews.
Share bugs, ideas, or general feedback.
Note: The current year is 2026. Use this when searching for recent documentation and best practices.
ce:plan-beta does the first planning pass. deepen-plan-beta is a second-pass confidence check.
Use this skill when the plan already exists and the question is not "Is this document clear?" but rather "Is this plan grounded enough for the complexity and risk involved?"
This skill does not turn plans into implementation scripts. It identifies weak sections, runs targeted research only for those sections, and strengthens the plan in place.
document-review and deepen-plan-beta are different:
document-review skill when the document needs clarity, simplification, completeness, or scope controldeepen-plan-beta when the document is structurally sound but still needs stronger rationale, sequencing, risk treatment, or system-wide thinkingUse the platform's question tool when available. When asking the user a question, prefer the platform's blocking question tool if one exists (AskUserQuestion in Claude Code, request_user_input in Codex, ask_user in Gemini). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
Ask one question at a time. Prefer a concise single-select choice when natural options exist.
<plan_path> #$ARGUMENTS </plan_path>
If the plan path above is empty:
docs/plans/ for recent filesDo not proceed until you have a valid plan file path.
Context & Research, Sources & References, and its origin document when present.ce:brainstorm.Read the plan file completely.
If the plan frontmatter includes an origin: path:
Determine the plan depth from the document:
Also build a risk profile. Treat these as high-risk signals:
Use this default:
If the plan already appears sufficiently grounded:
/ce:work or the document-review skillce:plan-beta StructureMap the plan into the current template. Look for these sections, or their nearest equivalents:
OverviewProblem FrameRequirements TraceScope BoundariesContext & ResearchKey Technical DecisionsOpen QuestionsHigh-Level Technical Design (optional overview — pseudo-code, DSL grammar, mermaid diagram, or data flow)Implementation Units (may include per-unit Technical design subsections)System-Wide ImpactRisks & DependenciesDocumentation / Operational NotesSources & ReferencesAlternative Approaches Considered, Success Metrics, Phased Delivery, Risk Analysis & Mitigation, and Operational / Rollout NotesIf the plan was written manually or uses different headings:
Also collect:
deepened: date if presentUse a checklist-first, risk-weighted scoring pass.
For each section, compute:
Key Technical Decisions, Implementation Units, System-Wide Impact, Risks & Dependencies, or Open Questions in Standard or Deep plansTreat a section as a candidate if:
Choose only the top 2-5 sections by score. If the user explicitly asked to deepen a lightweight plan, cap at 1-2 sections unless the topic is high-risk.
Example:
Key Technical Decisions section with 1 checklist trigger and the critical-section bonus scores 2 points and is a candidateRisks & Dependencies section with 1 checklist trigger in a high-risk migration plan also becomes a candidate because the risk bonus appliesIf the plan already has a deepened: date:
Use these triggers.
Requirements Trace
Context & Research / Sources & References
Key Technical Decisions
Open Questions
High-Level Technical Design (when present)
High-Level Technical Design (when absent) (Standard or Deep plans only)
Implementation Units
System-Wide Impact
Risks & Dependencies / Documentation / Operational Notes
Use the plan's own Context & Research and Sources & References as evidence. If those sections cite a pattern, learning, or risk that never affects decisions, implementation units, or verification, treat that as a confidence gap.
For each selected section, choose the smallest useful agent set. Do not run every agent. Use at most 1-3 agents per section and usually no more than 8 agents total.
Use fully-qualified agent names inside Task calls.
Requirements Trace / Open Questions classification
compound-engineering:workflow:spec-flow-analyzer for missing user flows, edge cases, and handoff gapscompound-engineering:research:repo-research-analyst (Scope: architecture, patterns) for repo-grounded patterns, conventions, and implementation reality checksContext & Research / Sources & References gaps
compound-engineering:research:learnings-researcher for institutional knowledge and past solved problemscompound-engineering:research:framework-docs-researcher for official framework or library behaviorcompound-engineering:research:best-practices-researcher for current external patterns and industry guidancecompound-engineering:research:git-history-analyzer only when historical rationale or prior art is materially missingKey Technical Decisions
compound-engineering:review:architecture-strategist for design integrity, boundaries, and architectural tradeoffscompound-engineering:research:framework-docs-researcher or compound-engineering:research:best-practices-researcher when the decision needs external grounding beyond repo evidenceHigh-Level Technical Design
compound-engineering:review:architecture-strategist for validating that the technical design accurately represents the intended approach and identifying gapscompound-engineering:research:repo-research-analyst (Scope: architecture, patterns) for grounding the technical design in existing repo patterns and conventionscompound-engineering:research:best-practices-researcher when the technical design involves a DSL, API surface, or pattern that benefits from external validationImplementation Units / Verification
compound-engineering:research:repo-research-analyst (Scope: patterns) for concrete file targets, patterns to follow, and repo-specific sequencing cluescompound-engineering:review:pattern-recognition-specialist for consistency, duplication risks, and alignment with existing patternscompound-engineering:workflow:spec-flow-analyzer when sequencing depends on user flow or handoff completenessSystem-Wide Impact
compound-engineering:review:architecture-strategist for cross-boundary effects, interface surfaces, and architectural knock-on impactcompound-engineering:review:performance-oracle for scalability, latency, throughput, and resource-risk analysiscompound-engineering:review:security-sentinel for auth, validation, exploit surfaces, and security boundary reviewcompound-engineering:review:data-integrity-guardian for migrations, persistent state safety, consistency, and data lifecycle risksRisks & Dependencies / Operational Notes
compound-engineering:review:security-sentinel for security, auth, privacy, and exploit riskcompound-engineering:review:data-integrity-guardian for persistent data safety, constraints, and transaction boundariescompound-engineering:review:data-migration-expert for migration realism, backfills, and production data transformation riskcompound-engineering:review:deployment-verification-agent for rollout checklists, rollback planning, and launch verificationcompound-engineering:review:performance-oracle for capacity, latency, and scaling concernsFor each selected section, pass:
Scope: architecture, patterns.) when the agent supports scoped invocationInstruct the agent to return:
Use the lightest mode that will work:
Signals that justify artifact-backed mode:
If artifact-backed mode is not clearly warranted, stay in direct mode.
Launch the selected agents in parallel using the execution mode chosen in Step 3.3. If the current platform does not support parallel dispatch, run them sequentially instead.
Prefer local repo and institutional evidence first. Use external research only when the gap cannot be closed responsibly from repo context or already-cited sources.
If a selected section can be improved by reading the origin document more carefully, do that before dispatching external agents.
Have each selected agent return its findings directly to the parent.
Keep the return payload focused:
If a direct-mode agent starts producing bulky or repetitive output, stop and switch the remaining research to artifact-backed mode instead of letting the parent context bloat.
Use a per-run scratch directory under .context/compound-engineering/deepen-plan-beta/, for example .context/compound-engineering/deepen-plan-beta/<run-id>/ or .context/compound-engineering/deepen-plan-beta/<plan-filename-stem>/.
Use the scratch directory only for the current deepening pass.
For each selected agent:
Prefer a compact markdown artifact unless machine-readable structure is clearly useful. Each artifact should contain:
Artifact rules:
Before synthesis:
If agent outputs conflict:
Strengthen only the selected sections. Keep the plan coherent and preserve its overall structure.
If artifact-backed mode was used:
Allowed changes:
Resolved During Planning and Deferred to Implementation when evidence supports the changedeepened: YYYY-MM-DD in frontmatter when the plan was substantively improvedDo not:
Research Insights subsections everywhereIf research reveals a product-level ambiguity that should change behavior or scope:
Open Questionsce:brainstorm if the gap is truly product-definingBefore writing:
Update the plan file in place by default.
If the user explicitly requests a separate file, append -deepened before .md, for example:
docs/plans/2026-03-15-001-feat-example-plan-deepened.mdIf artifact-backed mode was used and the user did not ask to inspect the scratch files:
If substantive changes were made, present next steps using the platform's blocking question tool when available (see Interaction Method). Otherwise, present numbered options in chat and wait for the user's reply before proceeding.
Question: "Plan deepened at [plan_path]. What would you like to do next?"
Options:
document-review skill - Improve the updated plan through structured document reviewce:work skill - Begin implementing the planBased on selection:
document-review skill -> Load the document-review skill with the plan pathce:work skill -> Call the ce:work skill with the plan pathIf no substantive changes were warranted:
document-review skill or /ce:work as the next step insteadNEVER CODE! Research, challenge, and strengthen the plan.