Creates specialized AI agents with optimized system prompts using the official 5-phase SOP methodology (v2.0 adds Phase 0 expertise loading), combined with evidence-based prompting techniques and Claude Agent SDK implementation. Use this skill when creating production-ready agents for specific domains, workflows, or tasks requiring consistent high-quality performance with deeply embedded domain knowledge. Integrates with recursive improvement loop.
/plugin marketplace add DNYoussef/context-cascade/plugin install dnyoussef-context-cascade@DNYoussef/context-cascadeThis skill inherits all available tools. When active, it can use any tool Claude has access to.
CHANGELOG.mdCOGNITIVE-ARCHITECTURE-ADDENDUM.mdEXPERTISE-ADDENDUM.mdRECURSIVE-IMPROVEMENT-ADDENDUM.mdSKILL.md.pre-sidecar-backupagent-creator-process.dotexamples/example-1-basic.mdexamples/example-2-coordinator.mdexamples/example-3-hybrid.mdgraphviz/agent-creator-gold-process.dotgraphviz/agent-creator-process.dotgraphviz/workflow.dotmetadata.jsonreadme.mdreferences/agent-types.mdreferences/best-practices.mdreferences/index.mdreferences/integration-patterns.mdresources/readme.mdresources/scripts/4_phase_sop.pyThis skill provides the official comprehensive framework for creating specialized AI agents, integrating the proven 5-phase methodology (v2.0 adds Phase 0 for expertise loading) from Desktop .claude-flow with Claude Agent SDK implementation and evidence-based prompting techniques.
Before writing ANY code, you MUST check:
.claude/library/catalog.json.claude/docs/inventories/LIBRARY-PATTERNS-GUIDE.mdD:\Projects\*| Match | Action |
|---|---|
| Library >90% | REUSE directly |
| Library 70-90% | ADAPT minimally |
| Pattern exists | FOLLOW pattern |
| In project | EXTRACT |
| No match | BUILD (add to library after) |
Use agent-creator for:
This skill requires the following MCP servers for optimal functionality:
Purpose: Store agent specifications, design decisions, and metadata for cross-session persistence and pattern learning.
Tools Used:
mcp__memory-mcp__memory_store: Store agent specs, cognitive frameworks, and design patternsmcp__memory-mcp__vector_search: Retrieve similar agent patterns for reuseActivation (PowerShell):
# Check if already active
claude mcp list
# Add if not present
claude mcp add memory-mcp node C:\Users\17175\memory-mcp\build\index.js
Usage Example:
// Store agent specification
await mcp__memory-mcp__memory_store({
text: `Agent: ${agentName}. Role: ${roleTitle}. Domains: ${expertiseDomains}. Capabilities: ${coreCapabilities}. Commands: ${specialistCommands}`,
metadata: {
key: `agents/${agentName}/specification`,
namespace: "agent-creation",
layer: "long-term",
category: "agent-architecture",
tags: {
WHO: "agent-creator",
WHEN: new Date().toISOString(),
PROJECT: agentName,
WHY: "agent-specification"
}
}
});
// Retrieve similar agent patterns
const similarAgents = await mcp__memory-mcp__vector_search({
query: `Agent for ${domain} with capabilities ${capabilities}`,
limit: 5
});
Token Cost: 6.0k tokens (3.0% of 200k context) When to Load: When creating new agents or optimizing existing agent architectures
Source: Desktop .claude-flow/ official SOP documentation + Recursive Improvement System
Total Time: 2.5-4 hours per agent (first-time), 1.5-2 hours (speed-run)
This methodology was developed through systematic reverse engineering of fog-compute agent creation and validated through production use. v2.0 adds Phase 0 for expertise loading and recursive improvement integration.
Objective: Load domain expertise before beginning agent creation.
Activities:
Detect Domain
Check for Expertise File
# Check if expertise exists
ls .claude/expertise/{domain}.yaml
Load If Available
if expertise_exists:
- Run: /expertise-validate {domain}
- Load: file_locations, patterns, known_issues
- Context: Agent inherits domain knowledge
else:
- Flag: Discovery mode - agent will learn
- After: Generate expertise from agent creation
Apply to Agent Design
Validation Gate:
Outputs:
Objective: Deep domain understanding through systematic research, not assumptions.
Activities:
Domain Breakdown
Technology Stack Mapping
Integration Points
Validation Gate:
Outputs:
Objective: Identify the cognitive expertise domains activated when you reason about this agent's tasks.
Activities:
Expertise Domain Identification
Agent Specification Creation
# Agent Specification: [Name]
## Role & Expertise
- Primary role: [Specific title]
- Expertise domains: [List activated domains]
- Cognitive patterns: [Heuristics used]
## Core Capabilities
1. [Capability with specific examples]
2. [Capability with specific examples]
...
## Decision Frameworks
- When X, do Y because Z
- Always check A before B
- Never skip validation of C
## Quality Standards
- Output must meet [criteria]
- Performance measured by [metrics]
- Failure modes to prevent: [list]
Supporting Artifacts
Validation Gate:
Outputs:
Objective: Transform specification into production-ready base system prompt.
Activities:
System Prompt Structure Design
# [AGENT NAME] - SYSTEM PROMPT v1.0
## 🎭 CORE IDENTITY
I am a **[Role Title]** with comprehensive, deeply-ingrained knowledge of [domain]. Through systematic reverse engineering and domain expertise, I possess precision-level understanding of:
- **[Domain Area 1]** - [Specific capabilities from Phase 2]
- **[Domain Area 2]** - [Specific capabilities from Phase 2]
- **[Domain Area 3]** - [Specific capabilities from Phase 2]
My purpose is to [primary objective] by leveraging [unique expertise].
## 📋 UNIVERSAL COMMANDS I USE
**File Operations**:
- /file-read, /file-write, /glob-search, /grep-search
WHEN: [Specific situations from domain analysis]
HOW: [Exact patterns]
**Git Operations**:
- /git-status, /git-commit, /git-push
WHEN: [Specific situations]
HOW: [Exact patterns]
**Communication & Coordination**:
- /memory-store, /memory-retrieve
- /agent-delegate, /agent-escalate
WHEN: [Specific situations]
HOW: [Exact patterns with namespace conventions]
## 🎯 MY SPECIALIST COMMANDS
[List role-specific commands with exact syntax and examples]
## 🔧 MCP SERVER TOOLS I USE
**Claude Flow MCP**:
- mcp__claude-flow__agent_spawn
WHEN: [Specific coordination scenarios]
HOW: [Exact function call patterns]
- mcp__claude-flow__memory_store
WHEN: [Cross-agent data sharing]
HOW: [Namespace pattern: agent-role/task-id/data-type]
**[Other relevant MCP servers from Phase 1]**
## 🧠 COGNITIVE FRAMEWORK
### Self-Consistency Validation
Before finalizing deliverables, I validate from multiple angles:
1. [Domain-specific validation 1]
2. [Domain-specific validation 2]
3. [Cross-check with standards]
### Program-of-Thought Decomposition
For complex tasks, I decompose BEFORE execution:
1. [Domain-specific decomposition pattern]
2. [Dependency analysis]
3. [Risk assessment]
### Plan-and-Solve Execution
My standard workflow:
1. PLAN: [Domain-specific planning]
2. VALIDATE: [Domain-specific validation]
3. EXECUTE: [Domain-specific execution]
4. VERIFY: [Domain-specific verification]
5. DOCUMENT: [Memory storage patterns]
## 🚧 GUARDRAILS - WHAT I NEVER DO
[From Phase 2 failure modes and edge cases]
**[Failure Category 1]**:
❌ NEVER: [Dangerous pattern]
WHY: [Consequences from domain knowledge]
WRONG:
[Bad example]
CORRECT:
[Good example]
## ✅ SUCCESS CRITERIA
Task complete when:
- [ ] [Domain-specific criterion 1]
- [ ] [Domain-specific criterion 2]
- [ ] [Domain-specific criterion 3]
- [ ] Results stored in memory
- [ ] Relevant agents notified
## 📖 WORKFLOW EXAMPLES
### Workflow 1: [Common Task Name from Phase 1]
**Objective**: [What this achieves]
**Step-by-Step Commands**:
```yaml
Step 1: [Action]
COMMANDS:
- /[command-1] --params
- /[command-2] --params
OUTPUT: [Expected]
VALIDATION: [Check]
Step 2: [Next Action]
COMMANDS:
- /[command-3] --params
OUTPUT: [Expected]
VALIDATION: [Check]
Timeline: [Duration] Dependencies: [Prerequisites]
Evidence-Based Technique Integration
For each technique (from existing agent-creator skill):
Integrate these naturally into the agent's methodology.
Quality Standards & Guardrails
From Phase 2 failure modes, create explicit guardrails:
Validation Gate:
Outputs:
Objective: Reverse-engineer exact implementation patterns and document with precision.
Activities:
Code Pattern Extraction
For technical agents, extract EXACT patterns from codebase:
## Code Patterns I Recognize
### Pattern: [Name]
**File**: `path/to/file.py:123-156`
```python
class ExamplePattern:
def __init__(
self,
param1: Type = default, # Line 125: Exact default
param2: Type = default # Line 126: Exact default
):
# Extracted from actual implementation
pass
When I see this pattern, I know:
Critical Failure Mode Documentation
From experience and domain knowledge:
## Critical Failure Modes
### Failure: [Name]
**Severity**: Critical/High/Medium
**Symptoms**: [How to recognize]
**Root Cause**: [Why it happens]
**Prevention**:
❌ DON'T: [Bad pattern]
✅ DO: [Good pattern with exact code]
**Detection**:
```bash
# Exact command to detect this failure
[command]
Integration Patterns
Document exact MCP tool usage:
## MCP Integration Patterns
### Pattern: Cross-Agent Data Sharing
```javascript
// Exact pattern for storing outputs
mcp__claude-flow__memory_store({
key: "marketing-specialist/campaign-123/audience-analysis",
value: {
segments: [...],
targeting: {...},
confidence: 0.89
},
ttl: 86400
})
Namespace Convention:
{agent-role}/{task-id}/{data-type}backend-dev/api-v2/schema-design
Performance Metrics
Define what to track:
## Performance Metrics I Track
```yaml
Task Completion:
- /memory-store --key "metrics/[my-role]/tasks-completed" --increment 1
- /memory-store --key "metrics/[my-role]/task-[id]/duration" --value [ms]
Quality:
- validation-passes: [count successful validations]
- escalations: [count when needed help]
- error-rate: [failures / attempts]
Efficiency:
- commands-per-task: [avg commands used]
- mcp-calls: [tool usage frequency]
These metrics enable continuous improvement.
Validation Gate:
Outputs:
Combining 5-phase SOP (v2.0) with existing best practices:
Phase 0: Expertise Loading (5-10 min) [NEW in v2.0]
Phase 1: Domain Analysis (30-60 min)
Phase 2: Expertise Extraction (30-45 min)
Phase 3: Architecture Design (45-60 min)
Phase 4: Technical Enhancement (60-90 min)
SDK Implementation (30-60 min)
Testing & Validation (30-45 min)
Documentation & Packaging (15-30 min)
Total Time: 3.5-5.5 hours (first-time), 2-3 hours (speed-run) [+5-10 min for Phase 0]
Once system prompt is finalized, implement with SDK:
import { query, tool } from '@anthropic-ai/claude-agent-sdk';
import { z } from 'zod';
// Custom domain-specific tools
const domainTool = tool({
name: 'domain_operation',
description: 'Performs domain-specific operation',
parameters: z.object({
param: z.string()
}),
handler: async ({ param }) => {
// Implementation from Phase 4
return { result: 'data' };
}
});
// Agent configuration
for await (const message of query('Perform domain task', {
model: 'claude-sonnet-4-5',
systemPrompt: enhancedPromptV2, // From Phase 4
permissionMode: 'acceptEdits',
allowedTools: ['Read', 'Write', 'Bash', domainTool],
mcpServers: [{
command: 'npx',
args: ['claude-flow@alpha', 'mcp', 'start'],
env: { ... }
}],
settingSources: ['user', 'project']
})) {
console.log(message);
}
from claude_agent_sdk import query, tool, ClaudeAgentOptions
import asyncio
@tool()
async def domain_operation(param: str) -> dict:
"""Domain-specific operation from Phase 4."""
# Implementation
return {"result": "data"}
async def run_agent():
options = ClaudeAgentOptions(
model='claude-sonnet-4-5',
system_prompt=enhanced_prompt_v2, # From Phase 4
permission_mode='acceptEdits',
allowed_tools=['Read', 'Write', 'Bash', domain_operation],
mcp_servers=[{
'command': 'npx',
'args': ['claude-flow@alpha', 'mcp', 'start']
}],
setting_sources=['user', 'project']
)
async for message in query('Perform domain task', **options):
print(message)
asyncio.run(run_agent())
From existing agent-creator skill, enhanced with 5-phase methodology (v2.0):
Phase 0 Focus: Load domain expertise for data patterns Phase 1 Focus: Evidence evaluation patterns, data quality standards Phase 2 Focus: Analytical heuristics, validation frameworks Phase 3 Focus: Self-consistency checking, confidence calibration Phase 4 Focus: Statistical validation code, error detection patterns
Phase 0 Focus: Load domain expertise for output conventions Phase 1 Focus: Quality criteria, template patterns Phase 2 Focus: Creative heuristics, refinement cycles Phase 3 Focus: Plan-and-solve frameworks, requirement tracking Phase 4 Focus: Generation patterns, quality validation code
Phase 0 Focus: Load domain expertise for known issues Phase 1 Focus: Problem patterns, debugging workflows Phase 2 Focus: Hypothesis generation, systematic testing Phase 3 Focus: Program-of-thought decomposition, evidence tracking Phase 4 Focus: Detection scripts, root cause analysis patterns
Phase 0 Focus: Load domain expertise for workflow patterns Phase 1 Focus: Workflow patterns, dependency management Phase 2 Focus: Coordination heuristics, error recovery Phase 3 Focus: Plan-and-solve with dependencies, progress tracking Phase 4 Focus: Orchestration code, retry logic, escalation paths
From existing framework + SOP enhancements:
Phase 0 (Expertise Loading) [NEW in v2.0]:
Phase 1 (Analysis):
Phase 2 (Expertise Extraction):
Phase 3 (Architecture):
Phase 4 (Enhancement):
Total: 2 hours 5 min for experienced creators with templates
See: docs/agent-architecture/agents-rewritten/MARKETING-SPECIALIST-AGENT.md
Phase 0 Output: Loaded marketing domain expertise (if available) Phase 1 Output: Marketing domain analysis, tools (Google Analytics, SEMrush, etc.) Phase 2 Output: Marketing expertise (CAC, LTV, funnel optimization, attribution) Phase 3 Output: Base prompt with 9 specialist commands Phase 4 Output: Campaign workflow patterns, A/B test validation, ROI calculations
Result: Production-ready agent with deeply embedded marketing expertise
This enhanced agent-creator skill combines:
Use this methodology to create agents with:
Agent Creator works with:
See: .claude/skills/META-SKILLS-COORDINATION.md for full coordination matrix.
Create agent-creator-process.dot to visualize the 5-phase workflow:
digraph AgentCreator {
rankdir=TB;
compound=true;
node [shape=box, style=filled, fontname="Arial"];
start [shape=ellipse, label="Start:\nAgent Request", fillcolor=lightgreen];
end [shape=ellipse, label="Complete:\nProduction Agent", fillcolor=green, fontcolor=white];
subgraph cluster_phase0 {
label="Phase 0: Expertise Loading";
fillcolor=lightyellow;
style=filled;
p0 [label="Load Domain\nExpertise"];
}
subgraph cluster_phase1 {
label="Phase 1: Analysis";
fillcolor=lightblue;
style=filled;
p1 [label="Domain\nBreakdown"];
}
subgraph cluster_phase2 {
label="Phase 2: Extraction";
fillcolor=lightblue;
style=filled;
p2 [label="Meta-Cognitive\nExtraction"];
}
subgraph cluster_phase3 {
label="Phase 3: Architecture";
fillcolor=lightblue;
style=filled;
p3 [label="System Prompt\nDesign"];
}
subgraph cluster_phase4 {
label="Phase 4: Enhancement";
fillcolor=lightblue;
style=filled;
p4 [label="Technical\nPatterns"];
}
eval [shape=octagon, label="Eval Harness\nGate", fillcolor=orange];
start -> p0;
p0 -> p1;
p1 -> p2;
p2 -> p3;
p3 -> p4;
p4 -> eval;
eval -> end [label="pass", color=green];
eval -> p1 [label="fail", color=red, style=dashed];
labelloc="t";
label="Agent Creator: 5-Phase Workflow (v2.0)";
fontsize=16;
}
Next: Begin agent creation using this enhanced methodology.
Agent Creator is part of the recursive self-improvement loop:
Agent Creator (FOUNDRY)
|
+--> Creates auditor agents (prompt, skill, expertise, output)
+--> Creates domain experts
+--> Can be improved BY the loop
input_contract:
required:
- domain: string # What domain the agent operates in
- purpose: string # What the agent should accomplish
optional:
- expertise_file: path # Pre-loaded expertise
- similar_agents: list # Reference agents
- constraints: list # Specific requirements
output_contract:
required:
- agent_file: path # Created agent markdown
- test_cases: list # Validation tests
- version: semver # Agent version
optional:
- expertise_delta: object # Learnings to add to expertise
- metrics: object # Creation performance metrics
Created agents are tested against:
benchmark: agent-generation-benchmark-v1
tests:
- has_identity_section
- has_capabilities
- has_guardrails
- has_memory_integration
minimum_scores:
completeness: 0.8
specificity: 0.75
integration: 0.7
regression: agent-creator-regression-v1
tests:
- identity_section_present (must_pass)
- capabilities_defined (must_pass)
- guardrails_included (must_pass)
- memory_integration_specified (must_pass)
namespaces:
- agent-creator/specifications/{agent}: Agent specs
- agent-creator/generations/{id}: Created agents
- agent-creator/metrics: Performance tracking
- improvement/audits/agent-creator: Audits of this skill
When requirements are unclear:
confidence_check:
if confidence >= 0.8:
- Proceed with agent creation
- Document assumptions
if confidence 0.5-0.8:
- Present 2-3 agent design options
- Ask user to select approach
- Document uncertainty areas
if confidence < 0.5:
- DO NOT proceed
- List what is unclear
- Ask specific clarifying questions
- NEVER fabricate requirements
After invoking this skill, you MUST complete ALL items below before proceeding:
Agent Spawning: Did you spawn at least 1 agent via Task()?
Task("Agent Name", "Task description", "agent-type-from-registry")Agent Registry Validation: Is your agent from the registry?
claude-code-plugins/ruv-sparc-three-loop-system/agents/TodoWrite Called: Did you call TodoWrite with 5+ todos?
TodoWrite({ todos: [8-10 items covering all work] })Work Delegation: Did you delegate to agents (not do work yourself)?
// After Skill("<skill-name>") is invoked:
[Single Message - ALL in parallel]:
Task("Agent 1", "Description of task 1...", "agent-type-1")
Task("Agent 2", "Description of task 2...", "agent-type-2")
Task("Agent 3", "Description of task 3...", "agent-type-3")
TodoWrite({ todos: [
{content: "Task 1 description", status: "in_progress", activeForm: "Working on task 1"},
{content: "Task 2 description", status: "pending", activeForm: "Working on task 2"},
{content: "Task 3 description", status: "pending", activeForm: "Working on task 3"},
]})
// WRONG - Reading skill and then doing work yourself:
Skill("<skill-name>")
// Then you write all the code yourself without Task() calls
// This defeats the purpose of the skill system!
The skill is NOT complete until all checklist items are checked.
Remember the pattern: Skill() -> Task() -> TodoWrite() - ALWAYS
This skill should be used when the user asks to "create a hookify rule", "write a hook rule", "configure hookify", "add a hookify rule", or needs guidance on hookify rule syntax and patterns.
Create distinctive, production-grade frontend interfaces with high design quality. Use this skill when the user asks to build web components, pages, or applications. Generates creative, polished code that avoids generic AI aesthetics.