From ai-model-research
Use when the user wants to find vision-capable / multimodal image-input models on OpenRouter. Triggers on phrases like "OpenRouter vision models", "image-capable models on OR", "models that accept images on OpenRouter", "multimodal models on OpenRouter", "OCR-capable OR models", "find a vision model on OpenRouter for <task>".
npx claudepluginhub danielrosehill/claude-code-plugins --plugin ai-model-researchThis skill uses the workspace's default tool permissions.
Filter the OpenRouter catalog to models that accept image input, then rank or summarize based on user criteria.
Mandates invoking relevant skills via tools before any response in coding sessions. Covers access, priorities, and adaptations for Claude Code, Copilot CLI, Gemini CLI.
Share bugs, ideas, or general feedback.
Filter the OpenRouter catalog to models that accept image input, then rank or summarize based on user criteria.
The user wants to discover models that can process images on OpenRouter — for OCR, image understanding, visual QA, document parsing, screenshot analysis, etc.
curl -s https://openrouter.ai/api/v1/models -H "Accept: application/json"
data[] where architecture.input_modalities includes "image".pricing.image field — many vision models charge a separate per-image fee on top of token costs. Surface this in output.pricing.prompt + pricing.imagecontext_length descendingimage in input but not output is for image understanding only — not generation. Image generation models also list image in output_modalities. If the user wants generation, route to that subset.Markdown table sorted by the user's criterion (default: cheapest combined cost). Include a count of total vision-capable models in the catalog.