site/docs/providers/openrouter.md
OpenRouter provides a unified interface for accessing various LLM APIs, including models from OpenAI, Meta, Perplexity, and others. It follows the OpenAI API format - see our OpenAI provider documentation for base API details.
OPENROUTER_API_KEY environment variable or specify apiKey in your configOpenRouter's catalog changes quickly. These are current popular and recent model IDs that work well as starting points. Context lengths are from the OpenRouter catalog at time of writing — check OpenRouter Models (or GET /api/v1/models) for live values.
| Model ID | Context (tokens) | Good for |
|---|---|---|
| openai/gpt-5.4 | 1,050,000 | Highest-quality general evaluation |
| anthropic/claude-opus-4.7 | 1,000,000 | Long-running agentic workflows |
| openai/gpt-5.4-mini | 400,000 | Fast, lower-cost GPT-5 workflows |
| anthropic/claude-haiku-4.5 | 200,000 | Lower-latency Claude runs |
| google/gemini-2.5-pro | 1,048,576 | Reasoning-heavy tasks |
| google/gemini-2.5-flash | 1,048,576 | Fast multimodal and general chat |
| meta-llama/llama-4-maverick | 1,048,576 | Popular open-weight frontier model |
| deepseek/deepseek-v3.2 | 163,840 | Cost-efficient reasoning and tools |
| mistralai/mistral-small-3.2-24b-instruct | 128,000 | Compact Mistral general use |
| qwen/qwen3-32b | 40,960 | Strong open multilingual model |
For the full catalog of 300+ models and current pricing, visit OpenRouter Models.
# yaml-language-server: $schema=https://promptfoo.dev/config-schema.json
providers:
- id: openrouter:openai/gpt-5.4
config:
temperature: 0.7
max_tokens: 1000
- id: openrouter:anthropic/claude-opus-4.7
config:
max_tokens: 2000
- id: openrouter:google/gemini-2.5-flash
config:
temperature: 0.7
max_tokens: 4000
If you route OpenRouter traffic through a proxy or OpenRouter-compatible gateway, set apiBaseUrl in the provider config. Precedence is config.apiBaseUrl → the hardcoded OpenRouter default (https://openrouter.ai/api/v1); the generic OpenAI OPENAI_API_BASE_URL / OPENAI_BASE_URL env fallbacks are not consulted for this provider.
The same pattern applies to apiKeyEnvar — set it to read your API key from a custom environment variable name (default OPENROUTER_API_KEY).
providers:
- id: openrouter:openai/gpt-5.4
config:
apiBaseUrl: https://proxy.example.com/openrouter/api/v1
apiKeyEnvar: MY_PROXY_KEY # optional: read the Bearer token from $MY_PROXY_KEY
Some models like Gemini 2.5 Pro include thinking tokens in their responses. You can control whether these are shown using the showThinking parameter:
providers:
- id: openrouter:google/gemini-2.5-pro
config:
showThinking: false # Hide thinking content from output (default: true)
When showThinking is true (default), the output includes thinking content:
Thinking: <reasoning process>
<actual response>