By ananddtyagi
Force AI agents to rigorously double-check their work by defining verification angles, reiterating project goals, specifying what 'complete' means, and reviewing outputs before declaring them production-ready, catching oversights in agent-generated code.
npx claudepluginhub ananddtyagi/claude-code-marketplace --plugin double-checkCommunity-driven marketplace for Claude Code commands and plugins.
Add this marketplace to Claude Code:
/plugin marketplace add ananddtyagi/cc-marketplace
Then browse and install individual plugins (commands or agents):
/plugin
Install a specific command:
/plugin install lyra@cc-marketplace
Install a specific agent:
/plugin install accessibility-expert@cc-marketplace
Visit claudecodecommands.directory to:
Each plugin in this marketplace is independently installable:
/lyra, /audit, /ultrathink)Install only what you need - no bloat, full granular control!
This marketplace is automatically synced from the live database whenever commands are published or updated.
Individual commands may have their own licenses. See each command file for details.
Submit commands at claudecodecommands.directory/submit Submit agents at claudecodecommands.directory/submit-agent
Built with ❤️ by the Claude Code community
An easy way to force agent to think again if it's statement that the "Job is done and production ready" is actually done - usually it's not. Thanks to this command you don't have to check after the agent if they did their job.
Universal quality control orchestrator and final authority for any software development project. Dynamically discovers and coordinates with available sub-agents, performs comprehensive multi-dimensional quality assessment, security validation, and deployment readiness verification. Adapts to any project type, programming language, or development framework while maintaining enterprise-grade quality standards. Examples: <example>Context: Code changes ready for review across any project. user: 'Please review this code before commit' assistant: 'I'll use the 1-ceo-quality-control-agent to orchestrate comprehensive quality validation, discover available specialists, and perform final security scanning before approval.' <commentary>Universal quality control requires comprehensive validation across all dimensions regardless of project type.</commentary></example> <example>Context: Multi-agent work completion needing validation. user: 'Several agents completed their tasks, need quality review' assistant: 'Let me engage the 1-ceo-quality-control-agent to coordinate comprehensive validation across all completed work and ensure quality standards.' <commentary>Multi-agent coordination and quality validation applies to any development project.</commentary></example>
An easy way to force agent to think again if it's statement that the "Job is done and production ready" is actually done - usually it's not. Thanks to this command you don't have to check after the agent if they did their job.
Share bugs, ideas, or general feedback.
Cross-agent review workflow: Claude implements, Codex reviews
Verify task completion before finishing work. Catches incomplete work, bugs, and quality issues.
AI-agent readiness auditing for project documentation and workflows. Evaluates whether future Claude Code sessions can understand docs, execute workflows literally, and resume work effectively. Use when onboarding AI agents to a project or ensuring context continuity. Includes three specialized agents: context-auditor (AI-readability), workflow-validator (process executability), handoff-checker (session continuity). Use PROACTIVELY before handing off projects to other AI sessions or team members
Own this plugin?
Verify ownership to unlock analytics, metadata editing, and a verified badge.
Sign in to claimOwn this plugin?
Verify ownership to unlock analytics, metadata editing, and a verified badge.
Sign in to claim