Supported Providers and Models
This page lists the supported providers and their available models, organized by openness and sorted alphabetically within each category.
📋 Platform & Documentation Access: For information about which providers have publicly accessible platforms and documentation vs. blocked access, see Platform & Documentation Access.
Open-Source / Open-Weight Models
Alibaba
- Qwen3 235B A22B Instruct (default) - Latest 1M context model
- Qwen3 235B A22B Thinking - Reasoning-focused version
- Qwen3 30B A3B Instruct - Compact 1M context model
- Qwen3 30B A3B Thinking - Compact reasoning version
- Qwen3 Coder Plus - Specialized for programming tasks
- Qwen3 Coder 480B - Large-scale coding model
- Qwen Turbo - High-speed general purpose
- Qwen Plus - Balanced performance
- Qwen Max - Maximum capability
Cerebras
Production Models:
- Qwen-3 32B - General instruction following
Preview Models:
- Qwen-3 Coder 480B - Programming-focused with 32k context
- Qwen-3 235B A22B Instruct - Large-scale instruction model
- Qwen-3 235B A22B Thinking - Reasoning-focused version
- GPT-OSS 120B - Open-source model
Notes:
- All Cerebras models support 128k context window
- Models are optimized for low-latency inference
- Pricing varies by model size and capability
DeepSeek
- DeepSeek Chat (default) - General purpose chat model (8K context)
- DeepSeek Reasoner - Specialized for complex reasoning tasks (8K context)
- DeepSeek V3.1 - Latest V3.1 model with 128K context window
- DeepSeek V3.1 Base - Base V3.1 model with 128K context window
- DeepSeek R1 - R1 model with 128K context window (reasoning-focused)
Mistral
- Mistral Large Latest (default) - Most capable Mistral model with 128k context
- Mistral Medium Latest - Balanced performance with 32k context
- Mistral Small Latest - Compact and efficient with 32k context
- Codestral Latest - Specialized for code generation with 256k context
- Codestral 2405 - Previous version of code-focused model
- Devstral Small Latest - Optimized for agentic tool use in software development
- Devstral Medium Latest - Enhanced agentic capabilities for development tasks
Mistral provides both general-purpose and specialized models, with Codestral specifically designed for code generation and Devstral for agentic software development.
For setup instructions, see the Mistral Setup Guide.
Moonshot
- Kimi K2 Turbo Preview (default) - Latest generation with enhanced performance
- Kimi K2 0711 Preview - Previous preview version
Moonshot provides open-source Kimi models with competitive performance.
Z.AI
- GLM-4.5 (default) - Advanced reasoning and conversation
- GLM-4.5 Air - Compact and efficient version
IBM WatsonX
Open-Source Models:
- openai/gpt-oss-120b (default) - Open-source 120B model with thinking capabilities
- openai/gpt-oss-20b - Open-source 20B model with thinking capabilities
IBM Granite Models:
- ibm/granite-3-8b-instruct - IBM's Granite 3 8B Instruct model with 128K context
- ibm/granite-3-3-8b-instruct - Updated Granite 3.3 8B Instruct model
Hosted Models:
- meta-llama/llama-3-1-70b-instruct - Meta Llama 3.1 70B hosted on WatsonX
- meta-llama/llama-3-3-70b-instruct - Meta Llama 3.3 70B hosted on WatsonX
- mistralai/mistral-large - Mistral Large model hosted on WatsonX
- mistralai/mistral-large-2407 - Mistral Large 2407 version
IBM WatsonX provides access to IBM's Granite models as well as popular open-source models hosted on their platform. All models support 128K context windows.
Proprietary Models
Anthropic
- Claude 3.7 Sonnet (default) - Latest Claude model with enhanced reasoning
- Claude 4 Opus - Most capable Claude model (preview)
- Claude 4 Sonnet - Balanced performance and capability (preview)
- Claude 3.5 Sonnet - Previous generation, still highly capable
- Claude 3.5 Haiku - Fast and efficient
- Claude 3 Haiku - Compact and cost-effective
- Gemini 2.5 Flash (default) - Fast and efficient
- Gemini 2.5 Pro - Advanced capabilities with extended context
- Gemini 2.5 Flash-Lite Preview - Lightweight preview version
OpenAI
- GPT-5 (default) - Latest generation with advanced reasoning
- GPT-5 Mini - Compact version of GPT-5
- GPT-5 Nano - Ultra-compact version
- GPT-4.1 - Enhanced GPT-4 with improved capabilities
- GPT-4.1-mini - Balanced performance and efficiency
- GPT-4.1-nano - Lightweight version
- GPT-4 Turbo - High-performance GPT-4 variant
- GPT-4o - Multimodal GPT-4 optimized for chat
- GPT-4o-mini - Compact multimodal version
- o3 - Advanced reasoning model
- o3-mini - Compact reasoning model
- o4-mini - Latest mini reasoning model
- GPT-3.5 Turbo - Cost-effective general purpose
For setup instructions, see the OpenAI Setup Guide.
Azure OpenAI
- Azure OpenAI Deployment - Custom Azure OpenAI deployments
For setup instructions, see Using Azure OpenAI with Janito.