From optimize-anything
Create or write an evaluator script for scoring text artifacts, prompts, or configs during gepa optimization. Use when asked to build, scaffold, or generate an evaluator, scoring function, or judge for optimize-anything.
npx claudepluginhub asragab/optimize-anythingThis skill uses the workspace's default tool permissions.
Generate an evaluator that scores candidate artifacts for optimization with gepa. Include diagnostic feedback so reflections can improve weak dimensions.
Guides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Automates semantic versioning and release workflow for Claude Code plugins: bumps versions in package.json, marketplace.json, plugin.json; verifies builds; creates git tags, GitHub releases, changelogs.
Generate an evaluator that scores candidate artifacts for optimization with gepa. Include diagnostic feedback so reflections can improve weak dimensions.
--evaluator-command) or HTTP POST body (--evaluator-url){"candidate": "<text>"}--dataset): {"candidate": "<text>", "example": {...}}score (float, usually in [0,1]), plus optional side-info fields.generate-evaluator now defaults to --evaluator-type judge.
response_format={"type": "json_object"} and includes dimension scores.--evaluator-type command
--evaluator-type http
--evaluator-type composite
score: 0.0.--evaluator-type judge|command|http|composite--model <litellm-model>: hardcodes judge model into judge/composite scripts.--dataset: generate dataset-aware templates that read example and show how to use it in scoring.--intake-json / --intake-file: embed rubric/quality dimensions.Generate a judge evaluator and test it:
# Generate
optimize-anything generate-evaluator seed.txt \
--objective "Score clarity and specificity" \
--model openai/gpt-4o-mini > eval_judge.py
# Test it
echo '{"candidate":"Your artifact text here"}' | python3 eval_judge.py
This returns JSON like:
{"score": 0.82, "reasoning": "Clear structure but lacks examples", "clarity": 0.9, "specificity": 0.7}
For dataset-aware evaluators:
optimize-anything generate-evaluator seed.txt \
--objective "Score correctness" \
--dataset examples.jsonl > eval_dataset.py
echo '{"candidate":"text","example":{"input":"q","expected":"a"}}' | python3 eval_dataset.py
score plus diagnostic fields.