From ouroboros
Delivers fast single-pass QA verdicts with 0-1 scores, PASS/REVISE/FAIL labels, and suggestions for code, documents, API responses, test outputs, or any artifact.
npx claudepluginhub q00/ouroboros --plugin ouroborosThis skill uses the workspace's default tool permissions.
Standalone quality assessment for any artifact — code, documents, API responses, test output, or custom content. Unlike `ooo evaluate` (3-stage formal verification pipeline), `ooo qa` is a fast single-pass verdict with actionable suggestions.
Evaluates AI-generated code quality using ICE Score (functional correctness, usefulness) and Code Judge metrics. Compares implementations, scores consistency, and lists inconsistencies against requirements.
Scores own output 0-10 across 5 task-appropriate dimensions pre-emit. Fixes and rescoring regressions below functional (5-6) band for complex code, docs, or specs.
Runs parallel specialized agents to verify implementations, run tests (unit/e2e/integration/perf/LLM), grade quality (0-10 scale), and suggest improvements. Use before merging.
Share bugs, ideas, or general feedback.
Standalone quality assessment for any artifact — code, documents, API responses, test output, or custom content. Unlike ooo evaluate (3-stage formal verification pipeline), ooo qa is a fast single-pass verdict with actionable suggestions.
ooo qa [file_path | artifact_text]
ooo qa # evaluate recent execution output
/ouroboros:qa [file_path | artifact_text] # plugin mode
Trigger keywords: "ooo qa", "qa check", "quality check"
The QA Judge evaluates an artifact against a quality bar and returns a structured verdict:
done (pass), continue (revise), escalate (fail)| Score Range | Verdict | Loop Action |
|---|---|---|
| >= 0.80 | PASS | done |
| 0.40 - 0.79 | REVISE | continue |
| < 0.40 | FAIL | escalate |
When the user invokes this skill:
This skill works in two modes. Determine which one before attempting any tool calls:
MCP mode — If ToolSearch is available, try loading the QA MCP tool:
ToolSearch query: "+ouroboros qa"
If found (typically named mcp__plugin_ouroboros_ouroboros__ouroboros_qa), proceed with QA Steps below.
Fallback mode — If ToolSearch is not available, or it finds no matching tool, skip directly to the Fallback section. This skill is designed to work without MCP setup.
Determine the artifact to evaluate:
Determine the quality bar:
Determine artifact type:
code — source code filestest_output — test results, CI outputdocument — specs, docs, READMEsapi_response — API responses, JSON payloadsscreenshot — visual artifactscustom — anything elseCall the ouroboros_qa MCP tool:
Tool: ouroboros_qa
Arguments:
artifact: <the content to evaluate>
quality_bar: <what 'pass' means>
artifact_type: "code" (or other type)
reference: <optional reference for comparison>
pass_threshold: 0.80 (adjustable)
seed_content: <seed YAML if available>
Present results clearly:
Next: Your artifact meets the quality bar. Proceed with confidence.Next: Address the suggestions above, then run ooo qa again to re-check.Next: Fundamental issues detected. Consider ooo interview to re-examine requirements, or ooo unstuck to challenge assumptions.For iterative usage, track the qa_session_id and iteration_history from the response meta:
qa_session_id and iteration_entry in metaqa_session_id and accumulated iteration_historypass or failIn fallback mode, generate a qa-<uuid4_short> session ID on the first run and maintain iteration count in conversation context to preserve the same iterative contract.
If the MCP server is not available, adopt the ouroboros:qa-judge agent role directly:
<project-root>/src/ouroboros/agents/qa-judge.md
(This is the same prompt used by the MCP QA tool, ensuring consistent verdicts.)QA Verdict [Iteration N]
========================
Session: qa-<id>
Score: X.XX / 1.00 [PASS/REVISE/FAIL]
Verdict: pass/revise/fail
Threshold: 0.80
Dimensions:
Correctness: X.XX
Completeness: X.XX
Quality: X.XX
Intent Alignment: X.XX
Domain-Specific: X.XX
Differences:
- <specific difference>
Suggestions:
- <actionable fix>
Reasoning: <1-3 sentence summary>
Loop Action: done/continue/escalate
User: ooo qa src/main.py
QA Verdict [Iteration 1]
============================================================
Session: qa-a1b2c3d4
Score: 0.72 / 1.00 [REVISE]
Verdict: revise
Threshold: 0.80
Dimensions:
Correctness: 0.85
Completeness: 0.60
Quality: 0.75
Intent Alignment: 0.80
Domain-Specific: 0.60
Differences:
- Missing error handling for network timeout in fetch_data()
- No input validation on user_id parameter
- Type hints missing on 3 public functions
Suggestions:
- Add try/except with TimeoutError in fetch_data() (line 42)
- Add isinstance check for user_id at function entry
- Add return type annotations to get_user(), fetch_data(), process_result()
Reasoning: Core logic is correct but lacks defensive programming
patterns expected for production code.
Loop Action: continue
Next: Address the suggestions above, then run `ooo qa` again to re-check.