From claude-skills
Improves existing Claude Code skills by modeling user intent, auditing effectiveness against a rubric, and proposing ranked enhancements like new features, UX gains, and efficiency fixes.
npx claudepluginhub gupsammy/claudest --plugin claude-skillsThis skill is limited to using the following tools:
Increase the effectiveness of an existing skill by modeling user intent, testing the skill
Optimizes a single existing Claude Code skill through task-based workflow: analyze structure, gather context, research best practices, identify/apply improvements, validate/review. For 'improve skill' or similar.
Improves existing Claude Code skills by fixing under/over-triggering, refining instructions, adding sub-skills, and evolving architecture based on feedback.
Refines and validates existing Claude Code skills for clarity, efficiency, and production readiness. Use for improving structure, best practices, token reduction, and production checks.
Share bugs, ideas, or general feedback.
Increase the effectiveness of an existing skill by modeling user intent, testing the skill against that intent through mental simulation and live doc validation, and proposing ranked improvements — new features, UX gains, accuracy fixes, efficiency wins.
This skill is about whether the skill accomplishes what users need. Structural correctness
(malformed frontmatter, missing fields, violated conventions) is repair-skill's domain — if
obvious structural issues are present, note them briefly, recommend repair-skill, and
continue with effectiveness analysis.
Read $ARGUMENTS as the path to a skill directory or SKILL.md file.
SKILL.md, then note which of references/, scripts/, examples/,
assets/ existIf the path is missing or ambiguous, use AskUserQuestion to resolve before proceeding.
Note any obvious structural issues in one sentence ("this description is first-person — recommend repair-skill for structural fixes") and move on. Do not run a full structural audit.
Phase 0 is complete when SKILL.md is loaded and the sibling directory inventory is noted.
Before analyzing, establish what the user wants. Use AskUserQuestion:
Regardless of the answer, also infer the skill's purpose from its description and body. State your understanding of what problem it solves and for whom — one sentence — before proceeding. This grounds the entire analysis in the correct frame.
If the user named a specific complaint: orient the analysis toward that area and scan for related issues in the same workflow region. If the user is unsure: run the full Phase 2 audit across all five sub-analyses.
Phase 1 is complete when user intent is established and the skill's purpose is stated.
Load ${CLAUDE_PLUGIN_ROOT}/skills/improve-skill/references/effectiveness-rubric.md before
starting. It contains the severity framework, improvement type definitions, and effort/impact
calibration criteria used in Phase 3.
Run all five sub-analyses. For each finding record: what the issue is, why it reduces effectiveness for the user, and the concrete improvement.
Walk through the skill as Claude executing it with a concrete representative user request. Choose an input that exercises the main workflow path — not an edge case, but the typical use.
For each phase of the skill, evaluate:
Document findings by type: stuck points, divergence points, dead ends, friction points.
Identify all factual claims in the skill that reference external standards: frontmatter field names, Claude tool names and behavior, API parameters, CLI flags, third-party service interfaces.
For each claim, verify against current documentation:
For Claude-specific claims (frontmatter options, tool names, model IDs):
Use Task tool with subagent_type=claude-code-guide — faster and more accurate than WebSearch.
For third-party claims (npm packages, APIs, external CLIs):
Use WebSearch or WebFetch against official documentation.
Flag drift between what the skill states and current reality. Severity: high if the claim produces broken output; medium if it produces outdated guidance; low if it's a naming change with no behavioral difference.
Given the skill's purpose, identify capabilities that are absent but would be high-value:
For each candidate: estimate implementation effort (one instruction change, new phase, or new script) and user value (rare edge case, common scenario, or blocks the skill in a key scenario).
Evaluate the skill's interaction design:
After walking the main path (2a), deliberately try to break the skill. Identify 3–5 adversarial inputs that test failure modes:
For each adversarial input, evaluate: does the skill detect the problem and surface a useful error, silently produce wrong output, or crash the workflow? Map findings to improvement types: missing error handling → NEW FEATURE, poor failure message → UX IMPROVEMENT, undetected bad state → ACCURACY FIX.
Phase 2 is complete when all five sub-analyses are finished and findings are recorded.
Load ${CLAUDE_PLUGIN_ROOT}/skills/improve-skill/references/effectiveness-report-template.md
for the output format before constructing the report. Reference
${CLAUDE_PLUGIN_ROOT}/skills/improve-skill/examples/sample-analysis.md to calibrate depth
and specificity if needed.
Present findings grouped by improvement type — users think in terms of outcomes, not audit
dimensions. Each entry must include: the sub-analysis code in brackets, what the gap is, why
it matters to the user, and the specific fix. Calibrate severity using the criteria in
references/effectiveness-rubric.md.
Ask: "Apply all improvements? Or select specific ones?"
Phase 3 is complete when the report is delivered and user selection is confirmed.
Apply confirmed items in order: new features → accuracy fixes → UX improvements → efficiency gains.
For each item:
After applying, briefly explain:
Validation: After delivering the explanation, re-read the modified SKILL.md in full and confirm: all selected improvements are present, no surrounding content was inadvertently altered, phases are still numbered with exit conditions, and all references point to files that exist.
Phase 4 is complete when all confirmed items are applied, the explanation is delivered, and the validation pass finds no integration failures.
After applying effectiveness improvements, invoke the skill-lint agent for a structural quality pass:
Use Task tool with subagent_type=claude-skills:skill-lint:
"Lint the skill at <path-to-skill-directory>. Auto-apply critical and major fixes, report
minor findings for user decision."
Wait for the agent to complete. If it auto-applied structural fixes, note them alongside the effectiveness changes from Phase 4. If it reports minor findings, present them to the user.
Phase 5 is complete when the lint agent returns and any user-selected minor fixes are applied.