From skill-creator
Create production-grade agent skills aligned with the 2026 AgentSkills.io spec and Anthropic best practices (2026). Also validates existing skills against the Intent Solutions 100-point rubric. Use when building, testing, validating, or optimizing Claude Code skills. Trigger with "/skill-creator", "create a skill", "validate my skill", or "check skill quality". Make sure to use this skill whenever creating a new skill, slash command, or agent capability.
npx claudepluginhub flight505/skill-forge --plugin skill-creatorThis skill is limited to using the following tools:
Creates complete, spec-compliant skill packages following AgentSkills.io and Anthropic standards.
agents/analyzer.mdagents/comparator.mdagents/grader.mdassets/eval_review.htmleval-viewer/generate_review.pyeval-viewer/viewer.htmlreferences/advanced-eval-workflow.mdreferences/anthropic-comparison.mdreferences/ard-template.mdreferences/creation-guide.mdreferences/errors-template.mdreferences/examples-template.mdreferences/frontmatter-spec.mdreferences/implementation-template.mdreferences/output-patterns.mdreferences/prd-template.mdreferences/schemas.mdreferences/source-of-truth.mdreferences/validation-rules.mdreferences/workflows.mdGuides Next.js Cache Components and Partial Prerendering (PPR): 'use cache' directives, cacheLife(), cacheTag(), revalidateTag() for caching, invalidation, static/dynamic optimization. Auto-activates on cacheComponents: true.
Guides building MCP servers enabling LLMs to interact with external services via tools. Covers best practices, TypeScript/Node (MCP SDK), Python (FastMCP).
Share bugs, ideas, or general feedback.
Creates complete, spec-compliant skill packages following AgentSkills.io and Anthropic standards. Supports both creation and validation workflows with 100-point marketplace grading.
Skill Creator solves the gap between writing ad-hoc agent skills and producing marketplace-ready
packages that score well on the Intent Solutions 100-point rubric. It enforces the 2026 spec
(top-level identity fields, ${CLAUDE_SKILL_DIR} paths, scored sections) and catches
contradictions that would cost marketplace points. Supports two modes: create new skills from
scratch with full validation, or grade/audit existing skills with actionable fix suggestions.
effort, maxTurns)validate-skill.py, aggregate_benchmark.py)~/.claude/skills/ or .claude/skills/)Determine user intent from their prompt:
Pay attention to context cues to understand the user's technical level. Skill creator is used by people across a wide range of familiarity — from first-time coders to senior engineers. In the default case:
If the current conversation already contains a workflow the user wants to capture (e.g., "turn this into a skill"), extract answers from the conversation history first — the tools used, the sequence of steps, corrections the user made, input/output formats observed. Confirm with the user before proceeding.
Ask the user with AskUserQuestion:
Skill Identity:
processing-pdfs, analyzing-data)Execution Model:
/name? Or background knowledge only?$ARGUMENTS substitution)context: fork for subagent execution)disable-model-invocation: true — prevents auto-activation, requires /name)Required Tools:
Bash(git:*), Bash(npm:*), etc.ServerName:tool_nameComplexity:
scripts/)references/)templates/)Location:
~/.claude/skills/<skill-name>/.claude/skills/<skill-name>/Before writing, determine:
Degrees of Freedom:
| Level | When to Use |
|---|---|
| High | Creative/open-ended tasks (analysis, writing) |
| Medium | Defined workflow, flexible content (most skills) |
| Low | Strict output format (compliance, API calls, configs) |
Think of it as narrow bridge vs open field: a deployment skill is a narrow bridge (one safe path, guard rails everywhere), while a writing skill is an open field (Claude roams freely within broad boundaries). Match constraint level to the task.
Workflow Pattern (see ${CLAUDE_SKILL_DIR}/references/workflows.md):
Output Pattern (see ${CLAUDE_SKILL_DIR}/references/output-patterns.md):
Create the skill directory and files:
mkdir -p {location}/{skill-name}
mkdir -p {location}/{skill-name}/scripts # if needed
mkdir -p {location}/{skill-name}/references # if needed
mkdir -p {location}/{skill-name}/templates # if needed
mkdir -p {location}/{skill-name}/assets # if needed
mkdir -p {location}/{skill-name}/evals # for eval-driven development
For detailed guidance on writing SKILL.md (frontmatter rules, description scoring, body guidelines, string substitutions, DCI syntax), creating supporting files, validation, testing, iteration, description optimization, and final reporting, see Creation Guide.
Key rules:
version, author, license, tags, compatible-with are TOP-LEVEL fields (not nested under metadata:)Bash(git:*) not bare Bashreferences/ if longerpython3 ${CLAUDE_SKILL_DIR}/scripts/validate-skill.py --grade {skill-dir}/SKILL.md to validateevals/evals.json with 3+ scenarios, iterate until all assertions passWhen the user wants to validate, grade, or audit an existing skill. For detailed steps (V1-V5), see Creation Guide.
~/.claude/skills/ or project .claude/skills/)python3 ${CLAUDE_SKILL_DIR}/scripts/validate-skill.py --grade {path}/SKILL.mdThe skill produces one of two outputs depending on mode:
scripts/, references/, templates/, assets/, and evals/ subdirectories, plus a creation summary report with validation grade and eval results.User: Create a skill called "code-review" that reviews code quality
Creates:
~/.claude/skills/code-review/
├── SKILL.md
└── evals/
└── evals.json
Frontmatter:
---
name: code-review
description: |
Make sure to use this skill whenever reviewing code for quality, security
vulnerabilities, and best practices. Use when doing code reviews, PR analysis,
or checking code quality. Trigger with "/code-review" or "review this code".
allowed-tools: "Read,Glob,Grep"
version: 1.0.0
author: Jeremy Longshore <jeremy@intentsolutions.io>
license: MIT
model: inherit
---
User: Create a skill that generates release notes from git history
Creates:
~/.claude/skills/generating-release-notes/
├── SKILL.md (argument-hint: "[version-tag]")
├── scripts/
│ └── parse-commits.py
├── references/
│ └── commit-conventions.md
├── templates/
│ └── release-template.md
└── evals/
└── evals.json
Uses $ARGUMENTS[0] for version tag.
Uses context: fork for isolated execution.
User: Grade my skill at ~/.claude/skills/code-review/SKILL.md
Runs: python3 ${CLAUDE_SKILL_DIR}/scripts/validate-skill.py --grade ~/.claude/skills/code-review/SKILL.md
Output:
Grade: B (84/100)
Improvements:
- Add "Trigger with" to description (+3 pts)
- Add ## Output section (+2 pts)
- Add ## Prerequisites section (+2 pts)
$ARGUMENTS, handle the empty caseBash, always scope itinherit, only override with good reason| Error | Cause | Solution |
|---|---|---|
| Name exists | Directory already present | Choose different name or confirm overwrite |
| Invalid name | Not kebab-case or >64 chars | Fix to lowercase-with-hyphens |
| Validation fails | Missing fields or anti-patterns | Run validator, fix reported issues |
| Resource missing | ${CLAUDE_SKILL_DIR}/ ref points to nonexistent file | Create the file or fix the reference |
| Undertriggering | Description too passive | Add "Make sure to use whenever..." phrasing |
| Eval failures | Skill not producing expected output | Iterate on instructions and re-test |
| Low grade | Missing scored sections or fields | Add Overview, Prerequisites, Output sections |
References: ${CLAUDE_SKILL_DIR}/references/
creation-guide.md — Detailed Steps 4-10 and Validation Workflow (V1-V5)source-of-truth.md — Canonical spec (AgentSkills.io, Anthropic docs, Lee Han Chung deep dive) | frontmatter-spec.md — Field reference | validation-rules.md — 100-point rubricworkflows.md — Workflow patterns | output-patterns.md — Output formats | schemas.md — JSON schemas (evals, grading, benchmarks)anthropic-comparison.md — Gap analysis | advanced-eval-workflow.md — Eval, iteration, optimization, platform notesAgents (read when spawning subagents): ${CLAUDE_SKILL_DIR}/agents/
grader.md — Assertion evaluation | comparator.md — Blind A/B comparison | analyzer.md — Benchmark analysisScripts: ${CLAUDE_SKILL_DIR}/scripts/
validate-skill.py — 100-point rubric grading | quick_validate.py — Lightweight validationaggregate_benchmark.py — Benchmark stats | run_eval.py — Trigger accuracy testingrun_loop.py — Description optimization loop | improve_description.py — LLM-powered rewritinggenerate_report.py — HTML reports | package_skill.py — .skill packaging | utils.py — Shared utilitiesEval Viewer: ${CLAUDE_SKILL_DIR}/eval-viewer/ — generate_review.py + viewer.html (interactive output comparison)
Assets: ${CLAUDE_SKILL_DIR}/assets/eval_review.html (trigger eval set editor)
Templates: ${CLAUDE_SKILL_DIR}/templates/skill-template.md (SKILL.md skeleton)
For advanced workflows (empirical eval, description optimization, blind comparison, packaging, platform notes), see Creation Guide and ${CLAUDE_SKILL_DIR}/references/advanced-eval-workflow.md.