Multi-model AI collaboration tools for Claude Code
npx claudepluginhub babywbx/superai-mcpMulti-model AI collaboration tools wrapping Codex, Gemini, and Claude CLIs as MCP tools with broadcast, chain, vote, debate, and quota checking
One MCP server to bridge them all: <br/> Unify Claude Code, Codex CLI, and Gemini CLI for seamless multi-model collaboration.<br/> Broadcast, chain, vote, debate — AI models working together.
English · 简体中文 · [Changelog][changelog] · [Issues][github-issues-link]
<!-- SHIELD GROUP -->[![][github-release-shield]][github-release-link] [![][python-shield]][python-link] [![][mcp-shield]][mcp-link] [![][github-license-shield]][github-license-link]<br/> [![][github-stars-shield]][github-stars-link] [![][github-forks-shield]][github-forks-link] [![][github-issues-shield]][github-issues-link] [![][github-contributors-shield]][github-contributors-link]<br/> [![][claude-shield]][claude-link] [![][codex-shield]][codex-link] [![][gemini-shield]][gemini-link]
Share SuperAI MCP
[![][share-x-shield]][share-x-link] [![][share-reddit-shield]][share-reddit-link] [![][share-telegram-shield]][share-telegram-link]
<sup>Multi-model collaboration, one MCP call away</sup>
</div> <details> <summary><kbd>Table of contents</kbd></summary>🔀 Multi-Model BroadcastSend the same prompt to Claude, Codex, and Gemini in parallel and get aggregated results in a single response. Perfect for multi-perspective code reviews, comparing model reasoning, or finding consensus across different AI engines.
models and overridesreview, refactor, explain, test, debug, optimize🔗 Chain PipelineBuild sequential multi-model pipelines where each step's output automatically flows into the next. Chain different models for tasks that benefit from staged processing — draft with one, refine with another, verify with a third.
<previous_output> tags between steps🗳️ Vote ConsensusRun candidates in parallel, then have a judge model pick the best answer. Get the most reliable output by leveraging the strengths of multiple models simultaneously.
💬 Multi-Round DebateStage an alternating debate between two models over multiple rounds. Each round sees the opponent's previous response. Ideal for exploring trade-offs, stress-testing solutions, or generating balanced analysis.
<opponent_response> tags⚡ Batch ProcessingRun N prompts against the same CLI concurrently in a single tool call. Bypasses stdio serialization bottleneck with server-side dispatch via asyncio.Semaphore. Up to 20 tasks with configurable concurrency.
🔁 Auto-Retry FallbackAutomatic cascading degradation on rate-limit, server error, or timeout. Retryable patterns include RESOURCE_EXHAUSTED, overloaded_error, 429, 5xx, timed out, etc.
| CLI | Fallback Strategy | Example |
|---|---|---|
| Gemini | Switch to flash model | pro → flash |
| Claude | Model downgrade | current → sonnet → haiku |
| Codex | Reduce reasoning effort | high → medium → low |
🗄️ Response CacheOpt-in LRU+TTL in-memory cache (use_cache=True). Identical requests (same cli+cd+prompt+model) return cached results instantly. Configure via SUPERAI_MCP_CACHE_TTL (default 300s) and SUPERAI_MCP_CACHE_MAXSIZE (default 128).
📡 Streaming OutputReal-time push of AI response chunks via MCP ctx.info() notifications (stream=True). Get immediate visibility into long-running tasks while the final complete response is still returned as the tool result.
🔍 Model Discovery & Validation