All models available through the Bankr LLM Gateway. Use bankr llm models to fetch the live list.
Claude (Anthropic)
| Model | Context | Input | Best For |
|---|
claude-opus-4.7 | 1M | text, image | Latest frontier, advanced reasoning |
claude-opus-4.6 | 1M | text, image | Advanced reasoning, analysis |
claude-opus-4.5 | 200K | text, image | Previous generation Opus |
claude-sonnet-4.6 | 1M | text, image | Balanced speed and quality |
claude-sonnet-4.5 | 1M | text, image | Previous generation Sonnet |
claude-haiku-4.5 | 200K | text, image | Fast, cost-effective |
Gemini (Google)
| Model | Context | Input | Best For |
|---|
gemini-3.1-pro | 1M | text, image | Latest frontier Gemini reasoning |
gemini-3.1-flash-lite | 1M | text, image | Ultra-fast, economical |
gemini-3-pro | — | — | Deprecated — use gemini-3.1-pro |
gemini-3-flash | 1M | text, image | Fast, high throughput |
gemini-2.5-pro | 1M | text, image | Long context, multimodal |
gemini-2.5-flash | 1M | text, image | Speed, high throughput |
Gemma (Google)
Open-weight Google models routed through Vertex AI (primary) with OpenRouter fallback.
| Model | Context | Input | Best For |
|---|
gemma-4-31b-it | 262K | text, image | Open-weight, multimodal |
gemma-4-26b-a4b-it | 262K | text, image | Open-weight, lightweight MoE |
GPT (OpenAI)
| Model | Context | Input | Best For |
|---|
gpt-5.5 | 1M | text, image | Latest frontier, native caching |
gpt-5.4 | 1M | text, image | Frontier OpenAI reasoning |
gpt-5.4-mini | 400K | text, image | Fast, economical |
gpt-5.4-nano | 400K | text, image | Ultra-fast, lowest cost |
gpt-5.2 | 400K | text | Previous generation reasoning |
gpt-5.2-codex | 400K | text | Code generation |
gpt-5-mini | 400K | text | Fast, economical |
gpt-5-nano | 400K | text | Ultra-fast, lowest cost |
Grok (xAI)
| Model | Context | Input | Best For |
|---|
grok-4.20 | 2M | text, image | xAI frontier reasoning |
grok-4.1-fast | 2M | text | Fast, long-context |
Kimi (Moonshot AI)
| Model | Context | Input | Best For |
|---|
kimi-k2.6 | 262K | text | Latest Moonshot reasoning |
kimi-k2.5 | 262K | text | Long-context reasoning |
Qwen (Alibaba)
| Model | Context | Input | Best For |
|---|
qwen3-coder | 262K | text | Code generation, debugging |
qwen3.5-plus | 1M | text | Long-context reasoning |
qwen3.5-flash | 1M | text | Fast, economical |
DeepSeek
| Model | Context | Input | Best For |
|---|
deepseek-v4-pro | 1M | text | Latest frontier, high-capacity reasoning |
deepseek-v4-flash | 1M | text | Fast inference, cost-effective |
deepseek-v3.2 | 163K | text | Previous generation reasoning |
GLM (Z.ai)
| Model | Context | Input | Best For |
|---|
glm-5.1 | 202K | text | Latest Z.ai reasoning |
glm-5 | 202K | text | Z.ai reasoning |
glm-5-turbo | 202K | text | Fast Z.ai |
MiniMax
| Model | Context | Input | Best For |
|---|
minimax-m2.7 | 204K | text | Latest MiniMax reasoning |
minimax-m2.7-highspeed | 204K | text | High-throughput MiniMax |
minimax-m2.5 | 204K | text | Long-context reasoning |