Evaluates and improves the exploration-cycle skills, prompts, routing, and artifact quality using baseline-first, one-hypothesis iteration loops with keep-discard decisions and experiment ledgers.
From exploration-cycle-pluginnpx claudepluginhub richfrem/agent-plugins-skills --plugin exploration-cycle-pluginThis skill is limited to using the following tools:
acceptance-criteria.mdevals/evals.jsonevals/results.tsvreferences/acceptance-criteria.mdreferences/architecture.mdreferences/spec-kitty-skill-optimizer-program.mdscripts/execute.pyGuides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Configures VPN and dedicated connections like Direct Connect, ExpressRoute, Interconnect for secure on-premises to AWS, Azure, GCP, OCI hybrid networking.
Ask for:
Confirm:
This skill implements autoresearch-style optimization for the exploration-cycle system. It uses a baseline-first iteration loop to improve skill prompts and logic.
Usage:
python3 ./scripts/execute.py \
--target ${plugins}/skills/user-story-capture/SKILL.md \
--eval-script ./scripts/eval_runner.py \
--goal "Improve Gherkin block accuracy" \
--iterations 3
For a concrete target-specific playbook, use references/spec-kitty-skill-optimizer-program.md when optimizing the Spec-Kitty agent/workflow files themselves.
The execute.py script follows a disciplined loop:
keep or discard.Always conclude execution with a Source Transparency Declaration explicitly listing what was queried to guarantee user trust: Sources Checked: [list] Sources Unavailable: [list]
./scripts/benchmarking/run_loop.py --results-dir evals/experiments for repeatable improvement loops.audit-plugin to verify the generated artifacts.