Skip to main content

Supported Models

All models available through the Bankr LLM Gateway. Use bankr llm models to fetch the live list.

Claude (Anthropic)

ModelContextInputBest For
claude-opus-4.71Mtext, imageLatest frontier, advanced reasoning
claude-opus-4.61Mtext, imageAdvanced reasoning, analysis
claude-opus-4.5200Ktext, imagePrevious generation Opus
claude-sonnet-4.61Mtext, imageBalanced speed and quality
claude-sonnet-4.51Mtext, imagePrevious generation Sonnet
claude-haiku-4.5200Ktext, imageFast, cost-effective

Gemini (Google)

ModelContextInputBest For
gemini-3.1-pro1Mtext, imageLatest frontier Gemini reasoning
gemini-3.1-flash-lite1Mtext, imageUltra-fast, economical
gemini-3-proDeprecated — use gemini-3.1-pro
gemini-3-flash1Mtext, imageFast, high throughput
gemini-2.5-pro1Mtext, imageLong context, multimodal
gemini-2.5-flash1Mtext, imageSpeed, high throughput

Gemma (Google)

Open-weight Google models routed through Vertex AI (primary) with OpenRouter fallback.

ModelContextInputBest For
gemma-4-31b-it262Ktext, imageOpen-weight, multimodal
gemma-4-26b-a4b-it262Ktext, imageOpen-weight, lightweight MoE

GPT (OpenAI)

ModelContextInputBest For
gpt-5.51Mtext, imageLatest frontier, native caching
gpt-5.41Mtext, imageFrontier OpenAI reasoning
gpt-5.4-mini400Ktext, imageFast, economical
gpt-5.4-nano400Ktext, imageUltra-fast, lowest cost
gpt-5.2400KtextPrevious generation reasoning
gpt-5.2-codex400KtextCode generation
gpt-5-mini400KtextFast, economical
gpt-5-nano400KtextUltra-fast, lowest cost

Grok (xAI)

ModelContextInputBest For
grok-4.202Mtext, imagexAI frontier reasoning
grok-4.1-fast2MtextFast, long-context

Kimi (Moonshot AI)

ModelContextInputBest For
kimi-k2.6262KtextLatest Moonshot reasoning
kimi-k2.5262KtextLong-context reasoning

Qwen (Alibaba)

ModelContextInputBest For
qwen3-coder262KtextCode generation, debugging
qwen3.5-plus1MtextLong-context reasoning
qwen3.5-flash1MtextFast, economical

DeepSeek

ModelContextInputBest For
deepseek-v4-pro1MtextLatest frontier, high-capacity reasoning
deepseek-v4-flash1MtextFast inference, cost-effective
deepseek-v3.2163KtextPrevious generation reasoning

GLM (Z.ai)

ModelContextInputBest For
glm-5.1202KtextLatest Z.ai reasoning
glm-5202KtextZ.ai reasoning
glm-5-turbo202KtextFast Z.ai

MiniMax

ModelContextInputBest For
minimax-m2.7204KtextLatest MiniMax reasoning
minimax-m2.7-highspeed204KtextHigh-throughput MiniMax
minimax-m2.5204KtextLong-context reasoning