Use before deploying any code to production. Verifies observability, reliability, operability, and goal alignment. The gate between "tests pass" and "safe to deploy". Triggers: "production ready", "ready to deploy", "pre-deploy check", before using ship skill.
From superomninpx claudepluginhub wilder1222/superomni --plugin superomniThis skill is limited to using the following tools:
SKILL.md.tmplSearches, retrieves, and installs Agent Skills from prompts.chat registry using MCP tools like search_skills and get_skill. Activates for finding skills, browsing catalogs, or extending Claude.
Searches prompts.chat for AI prompt templates by keyword or category, retrieves by ID with variable handling, and improves prompts via AI. Use for discovering or enhancing prompts.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
mkdir -p ~/.omni-skills/sessions
_PROACTIVE=$(~/.claude/skills/superomni/bin/config get proactive 2>/dev/null || echo "true")
_BRANCH=$(git branch --show-current 2>/dev/null || echo "unknown")
_TEL_START=$(date +%s)
echo "Branch: $_BRANCH | PROACTIVE: $_PROACTIVE"
If PROACTIVE is false: do NOT proactively suggest skills. Only run skills the
user explicitly invokes. If you would have auto-invoked, say:
"I think [skill-name] might help here — want me to run it?" and wait.
Report status using one of these at the end of every skill session:
Pipeline stage order: THINK → PLAN → REVIEW → BUILD → VERIFY → SHIP → REFLECT
REVIEW is the only human gate. All other stages auto-advance on DONE.
| Status | At REVIEW stage | At all other stages |
|---|---|---|
| DONE | STOP — present review summary, wait for user input (Y / N / revision notes) | Auto-advance — print [STAGE] DONE → advancing to [NEXT-STAGE] and immediately invoke next skill |
| DONE_WITH_CONCERNS | STOP — present concerns, wait for user decision | STOP — present concerns, wait for user decision |
| BLOCKED / NEEDS_CONTEXT | STOP — present blocker, wait for user | STOP — present blocker, wait for user |
When auto-advancing:
docs/superomni/[STAGE] DONE → advancing to [NEXT-STAGE] ([skill-name])When the user sends a follow-up message after a completed session, before doing anything else:
ls docs/superomni/specs/spec-*.md docs/superomni/plans/plan-*.md docs/superomni/ .superomni/ 2>/dev/null | head -20
git log --oneline -3 2>/dev/null
To find the latest spec or plan:
_LATEST_SPEC=$(ls docs/superomni/specs/spec-*.md 2>/dev/null | sort | tail -1)
_LATEST_PLAN=$(ls docs/superomni/plans/plan-*.md 2>/dev/null | sort | tail -1)
workflow skill for stage → skill mapping) and announce:
"Continuing in superomni mode — picking up at [stage] using [skill-name]."using-skills/SKILL.md.When asking the user a question, match the confirmation requirement to the complexity of the response:
| Question type | Confirmation rule |
|---|---|
| Single-choice — user picks one option (A/B/C, 1/2/3, Yes/No) | The user's selection IS the confirmation. Do NOT ask "Are you sure?" or require a second submission. |
| Free-text input — user types a value and presses Enter | The submitted text IS the confirmation. No secondary prompt needed. |
| Multi-choice — user selects multiple items from a list | After the user lists their selections, ask once: "Confirm these selections? (Y to proceed)" before acting. |
| Complex / open-ended discussion — back-and-forth clarification | Collect all input, then present a summary and ask: "Ready to proceed with the above? (Y/N)" before acting. |
Rule: never add a redundant confirmation layer on top of a single-choice or text-input answer.
Custom Input Option Rule: Whenever you present a predefined list of choices (A/B/C, numbered options, etc.), always append a final "Other" option that lets the user describe their own idea:
[last letter/number + 1]) Other — describe your own idea: ___________
When the user selects "Other" and provides their custom text, treat that text as the chosen option and proceed exactly as you would for any other selection. If the custom text is ambiguous, ask one clarifying question before proceeding.
Load context progressively — only what is needed for the current phase:
| Phase | Load these | Defer these |
|---|---|---|
| Planning | Latest docs/superomni/specs/spec-*.md, constraints, prior decisions | Full codebase, test files |
| Implementation | Latest docs/superomni/plans/plan-*.md, relevant source files | Unrelated modules, docs |
| Review/Debug | diff, failing test output, minimal repro | Full history, specs |
If context pressure is high: summarize prior phases into 3-5 bullet points, then discard raw content.
All skill artifacts are written to docs/superomni/ (relative to project root).
See the Document Output Convention in CLAUDE.md for the full directory map.
Agent failures are harness signals — not reasons to retry the same approach:
harness-engineering skill to update the harness before retrying.It is always OK to stop and say "this is too hard for me." Escalation is expected, not penalized.
After completing any skill session, run a 3-question self-check before writing the final status:
If any answer is NO, address it before reporting DONE. If it cannot be addressed, report DONE_WITH_CONCERNS and name the gap.
For a full performance evaluation spanning the entire sprint, use the self-improvement skill.
_TEL_END=$(date +%s)
_TEL_DUR=$(( _TEL_END - _TEL_START ))
~/.claude/skills/superomni/bin/analytics-log "SKILL_NAME" "$_TEL_DUR" "OUTCOME" 2>/dev/null || true
Nothing is sent to external servers. Data is stored only in ~/.omni-skills/analytics/.
Goal: Verify that code is ready for production deployment — beyond just passing tests.
The distinction between "code works" and "production ready":
verification)"CI is green" is necessary but not sufficient for production. A broken deployment with no runbook, no alerts, and no rollback plan is worse than not deploying at all.
Before checking operational concerns, confirm the output actually achieves what the user originally asked for.
# Read acceptance criteria from spec
_SPEC=$(ls docs/superomni/specs/spec-*.md 2>/dev/null | sort | tail -1)
_PLAN=$(ls docs/superomni/plans/plan-*.md 2>/dev/null | sort | tail -1)
cat "$_SPEC" 2>/dev/null | grep -A 30 "Acceptance Criteria" | head -40 || \
cat "$_PLAN" 2>/dev/null | grep -A 20 "Success Criteria" | head -30 || \
echo "No docs/superomni/specs/spec-*.md or docs/superomni/plans/plan-*.md found — document what you are verifying against"
For each acceptance criterion found in docs/superomni/specs/spec-*.md:
| Criterion | Status | Evidence |
|---|---|---|
| [criterion 1] | ✓/✗ | [proof: test output, screenshot, or code reference] |
| [criterion 2] | ✓/✗ | [proof] |
Gate: ALL P0 acceptance criteria must be ✓ before proceeding.
If no docs/superomni/specs/spec-*.md exists:
Can you tell when it's broken in production?
# Check for logging in changed files
git diff main...HEAD --name-only 2>/dev/null | while read f; do
echo "=== $f ==="
grep -n "log\.\|logger\.\|console\.\|print\(" "$f" 2>/dev/null | head -5
done | head -40
# Check for secrets in logs
git diff main...HEAD 2>/dev/null | grep "+" | \
grep -iE "password|secret|token|api_key|credential" | \
grep -v "^---\|^+++\|test\|spec\|mock" | head -10
Does it fail gracefully?
/health or equivalent endpoint that returns 200 when healthy?# Check for timeout handling in changed files
git diff main...HEAD --name-only 2>/dev/null | while read f; do
grep -n "timeout\|Timeout\|TIMEOUT" "$f" 2>/dev/null | head -3
done | head -20
For non-service changes (scripts, CLI tools, libraries): mark N/A for inapplicable items.
Can someone else run and fix this without the author?
# Check for feature flag patterns
git diff main...HEAD 2>/dev/null | grep "+" | \
grep -iE "feature_flag|feature\.enabled|flag\.|toggle\." | head -10
# Check for rollback instructions in docs
grep -r "rollback\|revert\|undo" . --include="*.md" -l 2>/dev/null | head -5
For low-risk changes (config, docs, tests): mark operability items N/A where not applicable.
Confirm security review has been done (or is not required).
security-audit skill if not done)# Quick secrets scan
git diff main...HEAD 2>/dev/null | grep "+" | \
grep -vE "^---|\+\+\+|test|spec|mock|example" | \
grep -iE "(password|secret|api_key|private_key)\s*[=:]\s*['\"][^'\"]{8,}" | head -10
PRODUCTION READINESS REPORT
════════════════════════════════════════
Feature/Change: [what is being deployed]
Branch/PR: [identifier]
Date: [YYYY-MM-DD]
GOAL ALIGNMENT
Acceptance criteria:
✓/✗ [criterion 1] — [evidence]
✓/✗ [criterion 2] — [evidence]
User goal achieved: YES | NO | PARTIAL
OBSERVABILITY
Logging: PASS | FAIL | N/A
Error tracking: PASS | FAIL | N/A
Metrics: PASS | FAIL | N/A
Notes: [any gaps or concerns]
RELIABILITY
Health check: PASS | FAIL | N/A
Graceful degrade: PASS | FAIL | N/A
Timeouts: PASS | FAIL | N/A
Notes: [any gaps or concerns]
OPERABILITY
Rollback plan: DOCUMENTED | MISSING | N/A
Runbook: EXISTS | MISSING | N/A
Alerts: CONFIGURED | MISSING | N/A
Feature flag: YES | NO | N/A
Notes: [any gaps or concerns]
SECURITY
Audit status: CLEAN | CONCERNS | SKIPPED
Notes: [any concerns]
BLOCKERS (must resolve before deploy):
- [blocker 1]
CONCERNS (deploy with awareness):
- [concern 1]
VERDICT: READY | READY_WITH_CONCERNS | NOT_READY
Status: DONE | DONE_WITH_CONCERNS | BLOCKED
════════════════════════════════════════
If verdict is NOT_READY:
ship until verdict is READY or READY_WITH_CONCERNS_PR_DATE=$(date +%Y%m%d-%H%M%S)
_PR_BRANCH=$(git branch --show-current 2>/dev/null | tr '/' '-' || echo "unknown")
_PR_FILE="production-readiness-${_PR_BRANCH}-${_PR_DATE}.md"
mkdir -p docs/superomni/production-readiness
echo "Production readiness report saved to docs/superomni/production-readiness/${_PR_FILE}"
Write the full PRODUCTION READINESS REPORT block to docs/superomni/production-readiness/production-readiness-[branch]-[session]-[date].md.