By evol-ai
Locally evaluate Claude Code/OpenClaw skills for quality across six dimensions and security risks using JS validators and LLM, score/compare versions, auto-evolve improvements via Ralph loops, merge git updates/rollbacks, scan portfolios, and generate reports—all stored under .skill-compass.
npx claudepluginhub evol-ai/skillcompass> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
- **Recommended model: Claude Opus 4.6** (`claude-opus-4-6`). Directed improvement requires understanding complex rubric feedback and generating precise, targeted edits. Weaker models may produce unfocused rewrites that fail to address the weakest dimension or introduce regressions in other dimensions.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
**🚀 Enhanced with Local Validators**: This command now uses local JavaScript validators for D1, D2, and D3 dimensions to significantly reduce token consumption while maintaining evaluation quality. Complex reasoning tasks (D4, D5, D6) continue to use LLM evaluation with local pre-analysis.
Triggered by the SessionStart hook (`hooks/scripts/session-tracker.js`), which compares the current SkillCompass version against the last recorded version. On first install, reinstall, or version change the hook injects a context message instructing Claude to load this file and follow it on the user's first interaction.
This command gives users a quick local inventory of installed skills and surfaces only high-signal issues. It supports two modes:
This command accepts free-form natural language and routes to the appropriate SkillCompass command. Also accessible as `/skill-compass`.
Unified entry point for managing skill suggestions and browsing all installed skills. Provides two views: suggestions (default) and all skills.
Generate a comprehensive report of all installed skills: quick health scan, context budget, portfolio overview, and quality summary.
> **Locale**: All templates in this spec are written in English. Detect the user's language from the session and translate user-facing text at display time per SKILL.md's Global UX Rules. Dimension labels: see the canonical table in SKILL.md.
Self-evolving skill engine for Claude Code. Creates, scores, repairs, and hardens skills autonomously through recursive improvement cycles.
Modifies files
Hook triggers on file write and edit operations
Runs pre-commands
Contains inline bash commands via ! syntax
Share bugs, ideas, or general feedback.
Evidence-based agent skills compiler with progressive capability tiers (Quick/Forge/Forge+/Deep).
Evaluate Agent Skill design quality against official specifications and best practices. Use when reviewing, auditing, or improving SKILL.md files and skill packages.
Analyze and optimize your Agent Skills (SKILL.md) using session data and research-backed static checks. Works with Claude Code, Codex, and any Agent Skills-compatible agent.
Tools for creating, auditing, and maintaining Claude Code skills. Includes /create-skill for scaffolding, /review-skill for quality checks, and /audit commands for bulk verification. Use when: building new skills, maintaining skill quality, or forking claude-skills repo.
Create new skills, improve existing skills, and measure skill performance. Use when users want to create a skill from scratch, update or optimize an existing skill, run evals to test a skill, or benchmark skill performance with variance analysis.
Bash prerequisite issue
Uses bash pre-commands but Bash not in allowed tools
Bash prerequisite issue
Uses bash pre-commands but Bash not in allowed tools
Share bugs, ideas, or general feedback.