Multi-model AI collaboration tools for Claude Code
npx claudepluginhub babywbx/superai-mcpMulti-model AI collaboration tools wrapping Codex, Gemini, and Claude CLIs as MCP tools with broadcast, chain, vote, debate, and quota checking
One MCP server to bridge them all:
Unify Claude Code, Codex CLI, and Gemini CLI for seamless multi-model collaboration.
Broadcast, chain, vote, debate — AI models working together.
English · 简体中文 · [Changelog][changelog] · [Issues][github-issues-link]
[![][github-release-shield]][github-release-link]
[![][python-shield]][python-link]
[![][mcp-shield]][mcp-link]
[![][github-license-shield]][github-license-link]
[![][github-stars-shield]][github-stars-link]
[![][github-forks-shield]][github-forks-link]
[![][github-issues-shield]][github-issues-link]
[![][github-contributors-shield]][github-contributors-link]
[![][claude-shield]][claude-link]
[![][codex-shield]][codex-link]
[![][gemini-shield]][gemini-link]
Share SuperAI MCP
[![][share-x-shield]][share-x-link] [![][share-reddit-shield]][share-reddit-link] [![][share-telegram-shield]][share-telegram-link]
Multi-model collaboration, one MCP call away
🔀 Multi-Model BroadcastSend the same prompt to Claude, Codex, and Gemini in parallel and get aggregated results in a single response. Perfect for multi-perspective code reviews, comparing model reasoning, or finding consensus across different AI engines.
models and overridesreview, refactor, explain, test, debug, optimize🔗 Chain PipelineBuild sequential multi-model pipelines where each step's output automatically flows into the next. Chain different models for tasks that benefit from staged processing — draft with one, refine with another, verify with a third.
<previous_output> tags between steps🗳️ Vote ConsensusRun candidates in parallel, then have a judge model pick the best answer. Get the most reliable output by leveraging the strengths of multiple models simultaneously.
💬 Multi-Round DebateStage an alternating debate between two models over multiple rounds. Each round sees the opponent's previous response. Ideal for exploring trade-offs, stress-testing solutions, or generating balanced analysis.
<opponent_response> tags⚡ Batch ProcessingRun N prompts against the same CLI concurrently in a single tool call. Bypasses stdio serialization bottleneck with server-side dispatch via asyncio.Semaphore. Up to 20 tasks with configurable concurrency.
🔁 Auto-Retry FallbackAutomatic cascading degradation on rate-limit, server error, or timeout. Retryable patterns include RESOURCE_EXHAUSTED, overloaded_error, 429, 5xx, timed out, etc.
| CLI | Fallback Strategy | Example |
|---|---|---|
| Gemini | Switch to flash model | pro → flash |
| Claude | Model downgrade | current → sonnet → haiku |
| Codex | Reduce reasoning effort | high → medium → low |
🗄️ Response CacheOpt-in LRU+TTL in-memory cache (use_cache=True). Identical requests (same cli+cd+prompt+model) return cached results instantly. Configure via SUPERAI_MCP_CACHE_TTL (default 300s) and SUPERAI_MCP_CACHE_MAXSIZE (default 128).
📡 Streaming OutputReal-time push of AI response chunks via MCP ctx.info() notifications (stream=True). Get immediate visibility into long-running tasks while the final complete response is still returned as the tool result.
🔍 Model Discovery & Validation