Skip to content

Supported Providers and Models

This page lists the supported providers and their available models, organized by openness and sorted alphabetically within each category.

📋 Platform & Documentation Access: For information about which providers have publicly accessible platforms and documentation vs. blocked access, see Platform & Documentation Access.

Open-Source / Open-Weight Models

Alibaba

  • Qwen3 235B A22B Instruct (default) - Latest 1M context model
  • Qwen3 235B A22B Thinking - Reasoning-focused version
  • Qwen3 30B A3B Instruct - Compact 1M context model
  • Qwen3 30B A3B Thinking - Compact reasoning version
  • Qwen3 Coder Plus - Specialized for programming tasks
  • Qwen3 Coder 480B - Large-scale coding model
  • Qwen Turbo - High-speed general purpose
  • Qwen Plus - Balanced performance
  • Qwen Max - Maximum capability

Cerebras

Production Models:

  • Qwen-3 32B - General instruction following

Preview Models:

  • Qwen-3 Coder 480B - Programming-focused with 32k context
  • Qwen-3 235B A22B Instruct - Large-scale instruction model
  • Qwen-3 235B A22B Thinking - Reasoning-focused version
  • GPT-OSS 120B - Open-source model

Notes:

  • All Cerebras models support 128k context window
  • Models are optimized for low-latency inference
  • Pricing varies by model size and capability

DeepSeek

  • DeepSeek Chat (default) - General purpose chat model (8K context)
  • DeepSeek Reasoner - Specialized for complex reasoning tasks (8K context)
  • DeepSeek V3.1 - Latest V3.1 model with 128K context window
  • DeepSeek V3.1 Base - Base V3.1 model with 128K context window
  • DeepSeek R1 - R1 model with 128K context window (reasoning-focused)

Mistral

  • Mistral Large Latest (default) - Most capable Mistral model with 128k context
  • Mistral Medium Latest - Balanced performance with 32k context
  • Mistral Small Latest - Compact and efficient with 32k context
  • Codestral Latest - Specialized for code generation with 256k context
  • Codestral 2405 - Previous version of code-focused model
  • Devstral Small Latest - Optimized for agentic tool use in software development
  • Devstral Medium Latest - Enhanced agentic capabilities for development tasks

Mistral provides both general-purpose and specialized models, with Codestral specifically designed for code generation and Devstral for agentic software development.

For setup instructions, see the Mistral Setup Guide.

Moonshot

  • Kimi K2 Turbo Preview (default) - Latest generation with enhanced performance
  • Kimi K2 0711 Preview - Previous preview version

Moonshot provides open-source Kimi models with competitive performance.

Z.AI

  • GLM-4.5 (default) - Advanced reasoning and conversation
  • GLM-4.5 Air - Compact and efficient version

IBM WatsonX

Open-Source Models:

  • openai/gpt-oss-120b (default) - Open-source 120B model with thinking capabilities
  • openai/gpt-oss-20b - Open-source 20B model with thinking capabilities

IBM Granite Models:

  • ibm/granite-3-8b-instruct - IBM's Granite 3 8B Instruct model with 128K context
  • ibm/granite-3-3-8b-instruct - Updated Granite 3.3 8B Instruct model

Hosted Models:

  • meta-llama/llama-3-1-70b-instruct - Meta Llama 3.1 70B hosted on WatsonX
  • meta-llama/llama-3-3-70b-instruct - Meta Llama 3.3 70B hosted on WatsonX
  • mistralai/mistral-large - Mistral Large model hosted on WatsonX
  • mistralai/mistral-large-2407 - Mistral Large 2407 version

IBM WatsonX provides access to IBM's Granite models as well as popular open-source models hosted on their platform. All models support 128K context windows.

Proprietary Models

Anthropic

  • Claude 3.7 Sonnet (default) - Latest Claude model with enhanced reasoning
  • Claude 4 Opus - Most capable Claude model (preview)
  • Claude 4 Sonnet - Balanced performance and capability (preview)
  • Claude 3.5 Sonnet - Previous generation, still highly capable
  • Claude 3.5 Haiku - Fast and efficient
  • Claude 3 Haiku - Compact and cost-effective

Google

  • Gemini 2.5 Flash (default) - Fast and efficient
  • Gemini 2.5 Pro - Advanced capabilities with extended context
  • Gemini 2.5 Flash-Lite Preview - Lightweight preview version

OpenAI

  • GPT-5 (default) - Latest generation with advanced reasoning
  • GPT-5 Mini - Compact version of GPT-5
  • GPT-5 Nano - Ultra-compact version
  • GPT-4.1 - Enhanced GPT-4 with improved capabilities
  • GPT-4.1-mini - Balanced performance and efficiency
  • GPT-4.1-nano - Lightweight version
  • GPT-4 Turbo - High-performance GPT-4 variant
  • GPT-4o - Multimodal GPT-4 optimized for chat
  • GPT-4o-mini - Compact multimodal version
  • o3 - Advanced reasoning model
  • o3-mini - Compact reasoning model
  • o4-mini - Latest mini reasoning model
  • GPT-3.5 Turbo - Cost-effective general purpose

For setup instructions, see the OpenAI Setup Guide.

Azure OpenAI

  • Azure OpenAI Deployment - Custom Azure OpenAI deployments

For setup instructions, see Using Azure OpenAI with Janito.