Applies TDD to skill creation/editing: run failing pressure scenarios first with subagents, write minimal SKILL.md to pass, refactor loopholes. Enforces 'no skill without failing test'.
npx claudepluginhub gadaalabs/claude-code-on-steroidsThis skill uses the workspace's default tool permissions.
**SCULPTOR** — *A sculptor does not add clay randomly — every choice is deliberate, tested against the form, and refined until it holds.*
Creates isolated Git worktrees for feature branches with prioritized directory selection, gitignore safety checks, auto project setup for Node/Python/Rust/Go, and baseline verification.
Executes implementation plans in current session by dispatching fresh subagents per independent task, with two-stage reviews: spec compliance then code quality.
Dispatches parallel agents to independently tackle 2+ tasks like separate test failures or subsystems without shared state or dependencies.
SCULPTOR — A sculptor does not add clay randomly — every choice is deliberate, tested against the form, and refined until it holds. When invoked: applies TDD to skill creation — write a failing test (pressure scenario) first, then write the skill to pass it, then refactor to close every loophole. No skill ships without watching an agent fail without it first.
Core principle: Creating skills IS Test-Driven Development applied to process documentation.
Announce at start: "Running SCULPTOR to write this skill with TDD discipline."
Required background: You must understand forge before using this skill. The RED-GREEN-REFACTOR cycle is identical — only the artifact changes.
NO SKILL WITHOUT A FAILING TEST FIRST
Applies to new skills AND edits to existing skills.
No exceptions:
| FORGE Concept | SCULPTOR Equivalent |
|---|---|
| Test case | Pressure scenario with subagent |
| Production code | SKILL.md |
| RED — test fails | Agent violates the rule without the skill |
| GREEN — test passes | Agent complies with the skill present |
| REFACTOR | Close loopholes while maintaining compliance |
| Write test first | Run baseline scenario BEFORE writing skill |
| Watch it fail | Document exact rationalizations the agent uses |
| Minimal code | Write skill addressing those specific violations only |
Create when:
Don't create for:
CLAUDE.md instead)---
name: skill-name
description: Use when [specific triggering conditions and symptoms]
type: process | domain | implementation
---
# Skill Name
## Overview
[Codename meaning + execution description]
Core principle in 1–2 sentences.
Announce at start: "Running [NAME] for [purpose]."
## Entry Point / When to Use
[Small flowchart ONLY if decision is non-obvious]
[Symptom-based bullet list]
## Core Pattern
[Steps, tables, code]
## Quick Reference
[Table for fast scanning]
## Red Flags
Never: ...
Always: ...
## Final Checklist
- [ ] Verifiable completion criterion
Skills are useless if they can't be found. Optimize for discovery.
Description = WHEN to use. Never WHAT it does.
# ❌ BAD: Summarizes workflow — Claude follows the description instead of reading the skill
description: Use when executing plans — dispatches subagent per task with review between tasks
# ✅ GOOD: Triggering conditions only
description: Use when executing implementation plans with independent tasks in the current session
Why this matters: When a description summarizes workflow, Claude takes it as a shortcut and skips reading the skill body. A description saying "code review between tasks" caused Claude to do ONE review, even though the skill showed TWO. Changing to triggering-conditions-only fixed it.
Format rules:
Include words Claude would search for:
"Hook timed out", "race condition", "ENOTEMPTY""flaky", "hanging", "zombie", "pollution""timeout/hang/freeze", "cleanup/teardown/afterEach"Target word counts:
Techniques:
patterns/ files (lazy-load)Load the patterns/persuasion-principles.md file for research-backed techniques on making skills behaviorally resilient under pressure.
Run a pressure scenario with a subagent without the skill. Document:
This is non-negotiable. You must see the failure before writing the skill.
Write the skill to address those specific rationalizations only. No speculative content.
Run the same scenario with the skill present. The agent should now comply.
Agent found a new rationalization? Add an explicit counter. Re-test. Repeat until bulletproof.
For discipline-enforcing skills, close loopholes explicitly:
State the rule AND forbid the workarounds:
❌ Write code before test? Delete it.
✅ Write code before test? Delete it. Start over.
No exceptions:
- Don't keep it as "reference"
- Don't "adapt" it while writing tests
- Don't look at it
- Delete means delete
Add the spirit-vs-letter block early:
Violating the letter of the rules is violating the spirit of the rules.
Build a rationalization table from your baseline test:
| Excuse | Reality |
|--------|---------|
| "Too simple to test" | Simple code breaks. Test takes 30 seconds. |
| "Tests after achieve the same goals" | Tests-after = "what does this do?" Tests-first = "what should this do?" |
RED Phase:
GREEN Phase:
name: uses letters, numbers, hyphens onlydescription: starts with "Use when...", triggering conditions only, third personREFACTOR Phase:
Quality Checks:
patterns/ filesNever:
Always:
Required background:
forge — the TDD cycle this skill adapts to documentationPairs with:
oracle — classify skill creation as a task before startingsentinel — verify the skill works before claiming it's done