From emasoft-orchestrator-agent
Use when verifying implementations. Trigger with verification, testing, or evidence requests.
npx claudepluginhub emasoft/emasoft-plugins --plugin emasoft-orchestrator-agentThis skill uses the workspace's default tool permissions.
This skill teaches evidence-based verification techniques for proving that code, systems, and operations work correctly.
README.mdreferences/automation-scripts.mdreferences/combining-patterns.mdreferences/cross-platform-support.mdreferences/docker-troubleshooting.mdreferences/end-to-end-testing.mdreferences/evidence-based-verification.mdreferences/evidence-format.mdreferences/exit-code-proof.mdreferences/github-integration.mdreferences/integration-verification.mdreferences/op-collect-evidence.mdreferences/op-format-verification-report.mdreferences/op-generate-test-report.mdreferences/op-notify-orchestrator.mdreferences/op-run-test-suite.mdreferences/op-update-github-issue.mdreferences/op-validate-evidence-record.mdreferences/op-verify-exit-code.mdreferences/test-report-format.mdProvides evidence-based verification patterns for code and systems, including exit code proofs, E2E testing, and integration checks. Generates pass/fail reports with reproducible evidence.
Verifies feature completion by writing automated tests against SPEC.md, running commands for fresh evidence, and confirming outputs per Iron Law of Verification.
Runs parallel specialized agents to verify implementations, run tests (unit/e2e/integration/perf/LLM), grade quality (0-10 scale), and suggest improvements. Use before merging.
Share bugs, ideas, or general feedback.
This skill teaches evidence-based verification techniques for proving that code, systems, and operations work correctly.
This skill teaches evidence-based verification techniques for proving that code, systems, and operations work correctly. Verification is not about hope or assumptions - it is about collecting measurable, reproducible evidence that a system behaves as intended.
The four core verification patterns are:
See Verification Principles below.
See Evidence-Based Verification:
See Exit Code Proof:
See End-to-End Testing:
See Combining Verification Patterns:
See Evidence Format Enforcement:
See Testing Protocol:
See Integration with GitHub Projects:
See Troubleshooting:
See Automation Scripts:
See Test Report Format:
def test_user_creation():
# Step 1: Define expected outcome
expected_email = "test@example.com"
# Step 2: Run the code
user = create_user(email=expected_email)
# Step 3: Collect evidence
actual_email = user.email
db_record = db.query(User).filter_by(id=user.id).first()
# Step 4: Compare evidence to expectation
assert actual_email == expected_email
assert db_record is not None
assert db_record.email == expected_email
# Step 5: Document results (test framework handles this)
# Run tests and capture exit code
pytest tests/ --cov=src
exit_code=$?
# Interpret result
if [ $exit_code -eq 0 ]; then
echo "SUCCESS: All tests passed"
else
echo "FAILURE: Tests failed with exit code $exit_code"
exit 1
fi
When verification tasks complete, provide results in this format:
| Field | Description | Example |
|---|---|---|
| Status | PASSED, FAILED, or ERROR | PASSED |
| Evidence Type | Type of verification performed | EXIT_CODE, FILE_CONTENT, TEST_RESULT |
| Evidence | Measurable proof collected | Exit code: 0, All 42 tests passed |
| Timestamp | When verification was performed | 2024-01-15T10:30:00Z |
| Details | Additional context or failure reasons | See test_results.json for details |
See Troubleshooting for complete solutions.
| Issue | Cause | Resolution |
|---|---|---|
| Tests pass locally, fail in CI | Environment differences | Check env vars, dependencies |
| Exit code 0 but process failed | Missing exit code in script | Add explicit sys.exit(1) on failure |
| Integration test timeout | Slow service startup | Increase timeout, add health checks |
| Flaky E2E test | Race conditions | Add explicit waits, retry logic |
Before learning specific verification patterns, understand these principles:
Do not assume code works. Do not say "this should work" or "probably works." Verify every claim with evidence.
Collect evidence that answers the question: "Does the system do what it is supposed to do?" Track:
Evidence is only valid if it can be reproduced. If you verify something once, you should be able to verify it again with the same result.
If something fails during verification, stop immediately and report the failure. Do not continue as if it succeeded.
Record what you verified, when you verified it, and what the results were. This documentation becomes your proof.
| Situation | Pattern | Reference |
|---|---|---|
| Prove a function returns correct value | Evidence-based verification | evidence-based-verification.md |
| Check if a script succeeded | Exit code proof | exit-code-proof.md |
| Verify user workflow works | E2E testing | end-to-end-testing.md |
| Test API calls database correctly | Integration verification | integration-verification.md |
| Build CI/CD pipeline | Exit code proof + E2E | combining-patterns.md |
| Cross-platform script | Platform-aware commands | cross-platform-support.md |
| Report to orchestrator | Evidence format | evidence-format.md |
| Run tests in isolation | Worktree testing | testing-protocol.md |
| Update GitHub issues | gh CLI integration | github-integration.md |
| Debug flaky tests | Troubleshooting guide | troubleshooting.md |
| Automate quality checks | Automation scripts | automation-scripts.md |
All verification scripts are in scripts/:
| Script | Purpose |
|---|---|
evidence_format.py | Evidence format validation and creation |
evidence_store.py | Evidence collection with deduplication |
consistency_verifier.py | File, git, URL, JSON verification |
with_server.py | Server orchestration for integration tests |
quality_pattern_detector.py | Anti-pattern detection |
scoring_framework.py | Weighted multi-dimension scoring |
comparison_analyzer.py | Gap analysis with baselines |
ab_test_calculator.py | Statistical hypothesis testing |
checklist_validator.py | Dependency-aware checklist validation |
traceability_validator.py | Requirements coverage validation |
| Code | Meaning | Action |
|---|---|---|
0 | Success | Continue workflow |
1 | General failure | Stop and report |
2 | Script validation failed | Fix script issues |
Verification is not optional. Every claim that code works must be backed by evidence. Use:
Combine these patterns to build confidence that your systems work correctly before deploying to production.
For detailed implementation of each pattern, see the reference files linked in the Table of Contents above.
Copy this checklist and track your progress: