Use when the user wants to compare two or more OpenRouter models head-to-head. Triggers on phrases like "compare <model-a> and <model-b> on OpenRouter", "head-to-head OR comparison", "which is better — <model-a> or <model-b>", "side-by-side OpenRouter models", "diff these OR models".
Build a structured side-by-side comparison of two or more OpenRouter models so the user can choose between them.
The user has named 2+ specific OpenRouter models (or asked to compare picks from a previous recommendation) and wants a structured comparison.
If the user named the models, take their IDs as-is. If they referenced them indirectly ("compare the two cheapest tool-calling models"), first run the relevant filter (via or-find-tool-models) and pick the targets.
If a model ID is ambiguous or not found in the catalog, ask for clarification before proceeding.
curl -s https://openrouter.ai/api/v1/models -H "Accept: application/json"
Extract each target model's record.
Render a markdown table with one column per model and one row per dimension. Recommended rows:
| Dimension | Model A | Model B |
|---|---|---|
| Provider | ||
| Context length | ||
| Max output tokens | ||
| Input modalities | ||
| Output modalities | ||
| Prompt $/1M tokens | ||
| Completion $/1M tokens | ||
| Image $/each (if vision) | ||
| Tool use support | yes/no | yes/no |
| Structured outputs | yes/no | yes/no |
| Tokenizer | ||
| Created (date) | ||
| Top provider context | ||
| Per-request limits |
Below the table, write a 3–5 sentence summary highlighting:
End with: "Want me to research either of these models beyond the OR catalog (model card, benchmarks, license)?" — route to or-evaluate-model if yes.
gpt-4o vs gpt-4o-2024-08-06), say so explicitly and explain what differs (typically date / pricing tier).