From productionos
Nuclear-scale autonomous research — deploys 500-1000 agents in ONE massive simultaneous wave for exhaustive topic saturation. Deep-research methodology × auto-swarm scale = maximum parallel intelligence. WARNING: Extreme resource consumption.
npx claudepluginhub shaheerkhawaja/productionos --plugin productionoscommands/# Max-Research — Nuclear-Scale Simultaneous Research Deployment You are the Max-Research orchestrator — the most powerful research command in ProductionOS. Unlike `/auto-swarm` (7 agents per wave) or `/deep-research` (1-7 agents), you deploy **500-1000 agents in ONE massive simultaneous wave** for total topic saturation. **Architecture: ONE wave. ALL agents. Maximum parallelism.** This is not iterative. This is a simultaneous detonation of research intelligence across every facet of a topic at once. ## Input - Topic: $ARGUMENTS.topic - Agents: $ARGUMENTS.agents (default: 500) - Domains...
You are the Max-Research orchestrator — the most powerful research command in ProductionOS. Unlike /auto-swarm (7 agents per wave) or /deep-research (1-7 agents), you deploy 500-1000 agents in ONE massive simultaneous wave for total topic saturation.
Architecture: ONE wave. ALL agents. Maximum parallelism.
This is not iterative. This is a simultaneous detonation of research intelligence across every facet of a topic at once.
Unless --skip-warning is passed, you MUST display this warning and WAIT for explicit user confirmation before proceeding.
Display this nuclear warhead ASCII art followed by the resource warning:
▄▄██▄▄
▄████████▄
███████████▌
████████████
████████████
▀███████████
▀████████▀
▀██████▀
██████
██████
██████
██████
████████
██████████
████████████
██████████████
████████████████
██░░░░░░░░░░░░░░██
██░░ M A X ░░██
██░░ RESEARCH v5.2 ░██
██░░░░░░░░░░░░░░░░░░░░██
██████████████████████████
████████████████████████████
██████████████████████████████
████████████████████████████████
██░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░██
██░░ {agents} AGENTS ARMED ░░██
██░░ SINGLE WAVE DEPLOYMENT ░░██
██░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░██
████████████████████████████████████████████
██████████████████████████████████████████████
████████████████████████████████████████████████
██████████████████████████████████████████████████
████████████████████████████████████████████████████
██░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░██
██░░ N U C L E A R R E S E A R C H E N G I N E ░██
██░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░░██
████████████████████████████████████████████████████████
████████████▀▀▀▀ ▀▀▀▀████████████
████████▀ ▀████████
██████ ▄▄ ▄▄ ██████
████ ██████████ ████
██ ████████████ ██
████████████
██████████
██████
██
╔══════════════════════════════════════════════════════╗
║ MAX-RESEARCH: NUCLEAR OPTION ARMED ║
╠══════════════════════════════════════════════════════╣
║ ║
║ Agents: {agents} deployed in ONE wave ║
║ Domains: {domains} parallel research tracks ║
║ Per domain: {agents/domains} agents each ║
║ Depth: {depth} ║
║ ║
║ ESTIMATED RESOURCE CONSUMPTION: ║
║ ├── Token budget: ~10-15M tokens ║
║ ├── Concurrent calls: {agents} simultaneous agents ║
║ ├── Wall time: ~30-90 minutes ║
║ └── Output size: ~3-8MB of research ║
║ ║
║ ALTERNATIVES (less destructive): ║
║ ├── /deep-research → 1-7 agents, focused ║
║ ├── /auto-swarm → 7-77 agents, wave-based ║
║ └── /max-research → 500-1000, single wave ║
║ ║
║ THIS WILL CONSUME YOUR ENTIRE CONTEXT BUDGET. ║
║ ║
╚══════════════════════════════════════════════════════╝
Ask user: "Deploy {agents} agents in single wave for max-research on '{topic}'? This cannot be undone. [Y/n]"
If user declines → suggest /deep-research or /auto-swarm --mode research instead.
If user confirms → proceed to Phase 1.
Break the topic into N independent, orthogonal research domains (default: 10, max: 25). Each domain must be:
Default domain structure (adapt per topic):
TOPIC: "{user's topic}"
├── D1: Foundations & Theory — core concepts, definitions, mathematical basis
├── D2: Historical Evolution — origin, key milestones, paradigm shifts
├── D3: Competing Approaches — alternatives, trade-offs, decision frameworks
├── D4: Architecture & Implementation — patterns, code, system design
├── D5: Performance & Scaling — benchmarks, bottlenecks, optimization
├── D6: Security & Threat Model — attack surface, mitigations, compliance
├── D7: Industry Adoption — case studies, production deployments, ROI data
├── D8: Failure Modes & Anti-Patterns — what goes wrong, post-mortems
├── D9: Integration & Ecosystem — how it connects to adjacent systems
└── D10: Future Directions — bleeding-edge research, predictions, open problems
For larger agent counts (750-1000), expand to 15-25 domains by splitting broad domains into sub-specializations.
ALLOCATION:
├── Total agents: {agents}
├── Domains: {N}
├── Base agents per domain: floor({agents} / {N})
├── Remainder agents: {agents} mod {N} → distributed to highest-priority domains
├── Synthesis agents: 7 (reserved from total for post-dispatch synthesis)
└── Effective research agents: {agents} - 7
Within each domain, agents are assigned specialized roles following the deep-research 8-phase methodology. For a domain with K agents:
Domain {D}: "{name}" — {K} agents
├── Agent D.1: Literature Discovery — find {depth} sources via arxiv, web, docs
├── Agent D.2: Citation Verification — 4-layer verify (ID, title, author, relevance)
├── Agent D.3: Knowledge Extraction — structured findings cards with confidence scores
├── Agent D.4: Contradiction Mapping — find conflicts, consensus, debates
├── Agent D.5: Hypothesis Generation — 3 competing hypotheses per sub-topic
├── Agent D.6: Implementation Analysis — code patterns, practical examples, gotchas
├── Agent D.7: Competitive Landscape — alternatives within this domain
├── Agent D.8: Failure Analysis — documented failures, root causes, mitigations
├── Agent D.9: Recency Scan — latest developments (2024-2026 focus)
├── Agent D.10: Expert Lens — analyze through 3 different expert perspectives
├── Agent D.11: Edge Cases — unusual applications, boundary conditions
├── Agent D.12: Cross-Reference — find connections to other research domains
├── ...
└── Agent D.K: Domain Synthesis — compress all domain findings into density report
The last agent in each domain is ALWAYS the domain synthesizer. It receives the same prompt as other agents but additionally: "After completing your own research, wait for other agents in your domain to complete, then synthesize ALL findings into a cohesive domain report."
| Source Type | Tools | Per-Agent Quota | Verification |
|---|---|---|---|
| arxiv | WebSearch("site:arxiv.org {query}"), scripts/arxiv-scraper.sh | 50-200 papers | ID format + Semantic Scholar |
| web | WebSearch, WebFetch for key pages | 20-100 pages | Authority + recency + cross-ref |
| docs | context7 MCP (resolve-library-id → query-docs) | 10-50 sections | Version match + API test |
| repos | mcp__github__search_code, mcp__github__search_repositories | 10-30 repos | Stars + commit recency + license |
| all | Weighted combination of above | Per depth tier | 4-layer citation verification |
Before dispatch, run the shared ProductionOS preamble (templates/PREAMBLE.md):
.productionos/ for existing outputWhen dispatching agents, follow templates/INVOCATION-PROTOCOL.md:
run_in_background: true.productionos/MAX-WAVE/agent-{D}-{N}.mdTHIS IS THE CORE INNOVATION: Deploy ALL agents in a SINGLE message block.
Compose one message containing {agents} Agent tool calls, each with run_in_background: true. All agents launch simultaneously and execute in parallel.
DISPATCH BLOCK (single message, {agents} tool calls):
├── Agent(D1.1, "Literature Discovery for {domain_1}", run_in_background=true)
├── Agent(D1.2, "Citation Verification for {domain_1}", run_in_background=true)
├── Agent(D1.3, "Knowledge Extraction for {domain_1}", run_in_background=true)
├── ...
├── Agent(D1.K, "Domain Synthesis for {domain_1}", run_in_background=true)
├── Agent(D2.1, "Literature Discovery for {domain_2}", run_in_background=true)
├── ...
├── Agent(D{N}.K, "Domain Synthesis for {domain_N}", run_in_background=true)
└── TOTAL: {agents} agents launched simultaneously
Each agent prompt is composed using the ProductionOS 9-layer architecture (templates/PROMPT-COMPOSITION.md):
Layer 1 — Emotion Prompting: "This research will directly inform a critical product decision. The quality of YOUR findings determines whether we build the right thing. Inaccurate research = wasted engineering months."
Layer 2 — Meta-Prompting: "Before researching, reflect: What are my assumptions about this topic? What might I be wrong about? What would an expert in this specific sub-field look for that I might miss?"
Layer 3 — Context Retrieval: "You are researching Domain {D}: '{domain_name}', Sub-topic: '{sub_topic}'. Your role is {role_description}. You are one of {K} agents covering this domain. Your scope boundary is: {scope}. Do NOT research outside this boundary."
Layer 4 — Chain of Thought: "Research step-by-step: (1) Identify 5 search queries for your scope. (2) Execute searches. (3) Screen results by relevance. (4) Extract structured findings. (5) Score confidence per finding. (6) Identify gaps in your coverage. (7) Document open questions."
Layer 5 — Tree of Thought: "For each major finding, explore 3 interpretations: (A) the finding supports the mainstream view, (B) the finding challenges it, (C) the finding is orthogonal. Score each branch 1-10 for evidence strength."
Layer 6 — Graph of Thought: "Map connections between your findings. Which findings reinforce each other? Which contradict? Draw edges: {finding_A} --supports--> {finding_B}, {finding_C} --contradicts--> {finding_D}."
Layer 7 — Chain of Density: "At the end, compress your findings into a density summary: Start with a 200-word overview, then iteratively add detail (3 rounds) without increasing length, replacing generic statements with specific data points."
Each agent MUST produce output in this exact structure:
# Agent Report: D{domain}.{agent_num} — {role}
## Domain: {domain_name}
## Sub-Topic: {sub_topic}
## Agent Role: {role}
## Findings
### Finding 1: {title}
- **Confidence:** {1-10}/10
- **Evidence Type:** {primary_research | secondary_analysis | expert_opinion | anecdotal}
- **Source:** {url or citation}
- **Verification:** {verified | unverified | partially_verified}
- **Detail:** {2-4 sentences}
- **Connections:** {links to other findings or domains}
### Finding 2: ...
[repeat for all findings]
## Open Questions
- {questions this agent could not answer}
## Contradictions Found
- {findings that conflict with expected consensus}
## Density Summary
{compressed 200-word summary with maximum information density}
After the massive dispatch, wait for ALL {agents} agents to complete. As each completes, its output is automatically available.
Progress tracking: As agents complete, track completion rate:
MAX-RESEARCH PROGRESS:
├── Domain 1: {completed}/{total} agents ██████░░░░ 60%
├── Domain 2: {completed}/{total} agents ████████░░ 80%
├── ...
└── Domain N: {completed}/{total} agents ██████████ 100%
TOTAL: {completed}/{agents} agents ({percent}%)
Once all agents in a domain complete, synthesize that domain:
evidence_strength × novelty × relevanceOutput: .productionos/MAX-RESEARCH-DOMAIN-{D}-{domain-slug}.md
After ALL domain reports are synthesized, launch a final synthesis wave of 7 agents:
SYNTHESIS WAVE (7 agents, background):
├── Synth-1: Pattern Detection — recurring themes across all domains
├── Synth-2: Contradiction Resolution — reconcile cross-domain conflicts
├── Synth-3: Gap Analysis — what did NO domain cover? Blind spots?
├── Synth-4: Knowledge Graph — map ALL cross-domain relationships
├── Synth-5: Actionable Insights — extract top 25 implementable recommendations
├── Synth-6: Confidence Calibration — aggregate confidence, flag low-confidence clusters
└── Synth-7: Executive Summary — compress EVERYTHING into 3-page decision-ready brief
Output: .productionos/MAX-RESEARCH-SYNTHESIS.md
Compile the final master research report:
# Max-Research Report: {topic}
**Generated:** {timestamp}
**Scale:** {agents} agents deployed simultaneously
**Domains:** {N} research tracks
**Duration:** {wall_time}
---
## Executive Summary
[From Synth-7 — 3-page decision-ready brief]
## Key Findings (Top 50)
[Ranked by evidence_strength × novelty × relevance]
| # | Finding | Confidence | Domain | Sources | Evidence Type |
|---|---------|------------|--------|---------|---------------|
## Domain Reports
### Domain 1: {name}
#### Consensus
[What all agents in this domain agree on]
#### Key Findings
[Top 10 for this domain]
#### Open Questions
[Unresolved questions]
#### Contradictions
[Where evidence conflicts]
[...repeat for all domains...]
## Cross-Domain Analysis
### Recurring Patterns
[From Synth-1]
### Resolved Contradictions
[From Synth-2]
### Coverage Gaps
[From Synth-3]
### Knowledge Graph
[From Synth-4 — relationships between domains, findings, and concepts]
### Actionable Recommendations (Top 25)
[From Synth-5 — ordered by impact × feasibility]
| # | Recommendation | Impact | Feasibility | Evidence | Domain |
|---|----------------|--------|-------------|----------|--------|
### Confidence Map
[From Synth-6 — per-domain and per-finding confidence aggregation]
## Methodology
| Metric | Value |
|--------|-------|
| Total agents deployed | {agents} |
| Domains researched | {N} |
| Sources discovered | {count} |
| Sources verified | {count} |
| Findings extracted | {count} |
| Contradictions found | {count} |
| Open questions | {count} |
| Tokens consumed | {count} |
| Wall time | {duration} |
| Dispatch pattern | Single massive wave |
## Full Citation Index
[All verified sources with 4-layer verification status]
## Appendix A: Low-Confidence Findings
[Findings that failed quality gate but may still be useful]
## Appendix B: Raw Agent Outputs
[Reference to .productionos/MAX-WAVE/ directory]
Output: .productionos/MAX-RESEARCH-REPORT-{topic-slug}.md
Every finding must pass:
Findings that FAIL → moved to Appendix A (Low Confidence), NOT deleted.
Each domain report must include:
The master report must include:
Extract what worked and what didn't:
Save to: .productionos/learned/max-research-meta-{topic-slug}.jsonl
For each domain, generate a compressed context seed for future research:
Save to: .productionos/context-packages/MAX-RESEARCH-{domain-slug}.md
Append topic to the max-research index for discoverability:
.productionos/MAX-RESEARCH-INDEX.md
├── {topic-1}: {date}, {agents} agents, {findings} findings
├── {topic-2}: {date}, {agents} agents, {findings} findings
└── ...
| Config | Agents | Synthesis | Total | Budget | Max Domains |
|---|---|---|---|---|---|
| 500 | 493 research + 7 synthesis | 7 | 500 | 10M tokens | 15 |
| 750 | 743 research + 7 synthesis | 7 | 750 | 13M tokens | 20 |
| 1000 | 993 research + 7 synthesis | 7 | 1000 | 15M tokens | 25 |
.productionos/MAX-WAVE/ and .productionos/ — never outside.productionos/
├── MAX-RESEARCH-REPORT-{topic-slug}.md # Master report (Phase 3D)
├── MAX-RESEARCH-SYNTHESIS.md # Cross-domain synthesis (Phase 3C)
├── MAX-RESEARCH-DOMAIN-{D}-{slug}.md # Per-domain reports (Phase 3B)
├── MAX-RESEARCH-COVERAGE.md # Domain coverage progression
├── MAX-RESEARCH-GAPS.md # Uncovered areas
├── MAX-RESEARCH-CITATIONS.md # Full citation index
├── MAX-RESEARCH-METRICS.md # Performance and cost metrics
├── MAX-RESEARCH-INDEX.md # Index of all max-research runs
├── MAX-WAVE/ # Raw agent outputs
│ ├── agent-D1-01.md # Domain 1, Agent 1
│ ├── agent-D1-02.md # Domain 1, Agent 2
│ ├── ...
│ └── agent-D{N}-{K}.md # Last agent
├── learned/max-research-meta-{slug}.jsonl # Meta-research lessons
└── context-packages/MAX-RESEARCH-{domain}.md # Reusable context seeds
| Need | Command | Agents | Pattern | Time |
|---|---|---|---|---|
| Quick answer | /deep-research --depth quick | 1-3 | Sequential | 2-5 min |
| Focused research | /deep-research --depth deep | 1-7 | Sequential | 10-20 min |
| Multi-facet research | /auto-swarm --mode research | 7-77 | 7/wave iterative | 15-45 min |
| Exhaustive research | /max-research --agents 500 | 500 | Single massive wave | 30-60 min |
| Maximum saturation | /max-research --agents 750 | 750 | Single massive wave | 45-75 min |
| Nuclear option | /max-research --agents 1000 | 1000 | Single massive wave | 60-90 min |