From ai-model-research
Use when the user wants to compare two or more OpenRouter models head-to-head. Triggers on phrases like "compare <model-a> and <model-b> on OpenRouter", "head-to-head OR comparison", "which is better — <model-a> or <model-b>", "side-by-side OpenRouter models", "diff these OR models".
npx claudepluginhub danielrosehill/claude-code-plugins --plugin ai-model-researchThis skill uses the workspace's default tool permissions.
Build a structured side-by-side comparison of two or more OpenRouter models so the user can choose between them.
Mandates invoking relevant skills via tools before any response in coding sessions. Covers access, priorities, and adaptations for Claude Code, Copilot CLI, Gemini CLI.
Share bugs, ideas, or general feedback.
Build a structured side-by-side comparison of two or more OpenRouter models so the user can choose between them.
The user has named 2+ specific OpenRouter models (or asked to compare picks from a previous recommendation) and wants a structured comparison.
If the user named the models, take their IDs as-is. If they referenced them indirectly ("compare the two cheapest tool-calling models"), first run the relevant filter (via or-find-tool-models) and pick the targets.
If a model ID is ambiguous or not found in the catalog, ask for clarification before proceeding.
curl -s https://openrouter.ai/api/v1/models -H "Accept: application/json"
Extract each target model's record.
Render a markdown table with one column per model and one row per dimension. Recommended rows:
| Dimension | Model A | Model B |
|---|---|---|
| Provider | ||
| Context length | ||
| Max output tokens | ||
| Input modalities | ||
| Output modalities | ||
| Prompt $/1M tokens | ||
| Completion $/1M tokens | ||
| Image $/each (if vision) | ||
| Tool use support | yes/no | yes/no |
| Structured outputs | yes/no | yes/no |
| Tokenizer | ||
| Created (date) | ||
| Top provider context | ||
| Per-request limits |
Below the table, write a 3–5 sentence summary highlighting:
End with: "Want me to research either of these models beyond the OR catalog (model card, benchmarks, license)?" — route to or-evaluate-model if yes.
gpt-4o vs gpt-4o-2024-08-06), say so explicitly and explain what differs (typically date / pricing tier).