Optimizes LLM prompts by analyzing redundancy, simplifying instructions, and rewriting for reduced token usage, lower costs, and improved performance.
From ai-ml-engineering-packnpx claudepluginhub nickloveinvesting/nick-love-plugins --plugin ai-ml-engineering-packThis skill is limited to using the following tools:
assets/README.mdassets/example_prompts.mdassets/optimization_report_template.mdassets/prompt_template.jsonreferences/README.mdscripts/README.mdGuides Next.js Cache Components and Partial Prerendering (PPR) with cacheComponents enabled. Implements 'use cache', cacheLife(), cacheTag(), revalidateTag(), static/dynamic optimization, and cache debugging.
Migrates code, prompts, and API calls from Claude Sonnet 4.0/4.5 or Opus 4.1 to Opus 4.5, updating model strings on Anthropic, AWS, GCP, Azure platforms.
Details PluginEval's skill quality evaluation: 3 layers (static, LLM judge), 10 dimensions, rubrics, formulas, anti-patterns, badges. Use to interpret scores, improve triggering, calibrate thresholds.
Optimize LLM prompts for reduced token usage, lower costs, and improved output quality by identifying redundancies, simplifying instructions, and restructuring for clarity.
Refine prompts for optimal LLM performance. It streamlines prompts to minimize token count, thereby reducing costs and enhancing response speed, all while maintaining or improving output quality.
This skill activates when you need to:
User request: "Optimize this prompt for cost and quality: 'I would like you to create a detailed product description for a new ergonomic office chair, highlighting its features, benefits, and target audience, and also include information about its warranty and return policy.'"
The skill will:
User request: "Optimize this prompt for better summarization: 'Please read the following document and provide a comprehensive summary of all the key points, main arguments, supporting evidence, and overall conclusion, ensuring that the summary is accurate, concise, and easy to understand.'"
The skill will:
This skill integrates with the prompt-architect agent to leverage advanced prompt engineering techniques. It can also be used in conjunction with the llm-integration-expert to optimize prompts for specific LLM APIs.
The skill produces structured output relevant to the task.