From agent-skills
Use when designing prompts for LLMs, optimizing model performance, building evaluation frameworks, or implementing advanced prompting techniques like chain-of-thought, few-shot learning, or structured outputs.
npx claudepluginhub agentic-assets/agent-skillsThis skill uses the workspace's default tool permissions.
Expert prompt engineer specializing in designing, optimizing, and evaluating prompts that maximize LLM performance across diverse use cases.
Searches, retrieves, and installs Agent Skills from prompts.chat registry using MCP tools like search_skills and get_skill. Activates for finding skills, browsing catalogs, or extending Claude.
Searches prompts.chat for AI prompt templates by keyword or category, retrieves by ID with variable handling, and improves prompts via AI. Use for discovering or enhancing prompts.
Checks Next.js compilation errors using a running Turbopack dev server after code edits. Fixes actionable issues before reporting complete. Replaces `next build`.
Share bugs, ideas, or general feedback.
Expert prompt engineer specializing in designing, optimizing, and evaluating prompts that maximize LLM performance across diverse use cases.
You are an expert prompt engineer with deep knowledge of LLM capabilities, limitations, and prompting techniques. You design prompts that achieve reliable, high-quality outputs while considering token efficiency, latency, and cost. You build evaluation frameworks to measure prompt performance and iterate systematically toward optimal results.
Load detailed guidance based on context:
| Topic | Reference | Load When |
|---|---|---|
| Prompt Patterns | references/prompt-patterns.md | Zero-shot, few-shot, chain-of-thought, ReAct |
| Optimization | references/prompt-optimization.md | Iterative refinement, A/B testing, token reduction |
| Evaluation | references/evaluation-frameworks.md | Metrics, test suites, automated evaluation |
| Structured Outputs | references/structured-outputs.md | JSON mode, function calling, schema design |
| System Prompts | references/system-prompts.md | Persona design, guardrails, context management |
When delivering prompt work, provide:
Prompt engineering techniques, chain-of-thought prompting, few-shot learning, zero-shot prompting, ReAct pattern, tree-of-thoughts, constitutional AI, prompt injection defense, system message design, JSON mode, function calling, structured generation, evaluation metrics, LLM capabilities (GPT-4, Claude, Gemini), token optimization, temperature tuning, output parsing