Visualize whether skills, rules, and agent definitions are actually followed — auto-generates scenarios at 3 prompt strictness levels, runs agents, classifies behavioral sequences, and reports compliance rates with full tool call timelines
From everything-claude-codenpx claudepluginhub binzetss/mobile-hvgllocalThis skill uses the workspace's default tool permissions.
fixtures/tdd_spec.yamlprompts/classifier.mdprompts/scenario_generator.mdprompts/spec_generator.mdDesigns and optimizes AI agent action spaces, tool definitions, observation formats, error recovery, and context for higher task completion rates.
Enables AI agents to execute x402 payments with per-task budgets, spending controls, and non-custodial wallets via MCP tools. Use when agents pay for APIs, services, or other agents.
Compares coding agents like Claude Code and Aider on custom YAML-defined codebase tasks using git worktrees, measuring pass rate, cost, time, and consistency.
Measures whether coding agents actually follow skills, rules, or agent definitions by:
claude -p and capturing tool call traces via stream-jsonskills/*/SKILL.md): Workflow skills like search-first, TDD guidesrules/common/*.md): Mandatory rules like testing.md, security.md, git-workflow.mdagents/*.md): Whether an agent gets invoked when expected (internal workflow verification not yet supported)/skill-comply <path># Full run
uv run python -m scripts.run ~/.claude/rules/common/testing.md
# Dry run (no cost, spec + scenarios only)
uv run python -m scripts.run --dry-run ~/.claude/skills/search-first/SKILL.md
# Custom models
uv run python -m scripts.run --gen-model haiku --model sonnet <path>
Measures whether a skill/rule is followed even when the prompt doesn't explicitly support it.
Reports are self-contained and include:
For users familiar with hooks, reports also include hook promotion recommendations for steps with low compliance. This is informational — the main value is the compliance visibility itself.