Supported LLMs

Chat and Prompts

Cody supports a variety of cutting-edge large language models for use in chat and prompts, allowing you to select the best model for your use case.

Newer versions of Sourcegraph Enterprise, starting from v5.6, it will be even easier to add support for new models and providers, see Model Configuration for more information.
ProviderModelFreeProEnterprise
OpenAIGPT-4 Turbo-
OpenAIGPT-4o-
OpenAIGPT-4o-mini
OpenAIo3-mini-medium (experimental)
OpenAIo3-mini-high (experimental)--
OpenAIo3-
OpenAIo4-mini
OpenAIGPT-4.1-
OpenAIGPT-4.1-mini
OpenAIGPT-4.1-nano
AnthropicClaude 3.5 Haiku
AnthropicClaude 3.5 Sonnet
AnthropicClaude 3.7 Sonnet-
AnthropicClaude Sonnet 4
AnthropicClaude Sonnet 4 w/Thinking-
AnthropicClaude Opus 4--
AnthropicClaude Opus 4 w/Thinking--
GoogleGemini 1.5 Pro✅ (beta)
GoogleGemini 2.0 Flash
GoogleGemini 2.0 Flash
GoogleGemini 2.5 Pro Preview-
GoogleGemini 2.5 Flash Preview (experimental)
To use Claude 3 Sonnet models with Cody Enterprise, make sure you've upgraded your Sourcegraph instance to the latest version.

Claude 3.7 and 4 Sonnet

Claude 3.7 and 4 Sonnet have two variants; the base version, and the extended thinking version which supports deep reasoning and fast, responsive edit workflows. Cody enables using both, and lets the user select which to use in the model dropdown selector, so the user can choose whether to use extended thinkig depending on their work task.

Claude 4 models support is available with Sourcegraph versions v6.4+ and v6.3.4167. It also depends on the deployement type and whether it's thinking or not.

Claude 3.7 and 4 via Google Vertex, via AWS Bedrock

Starting in Sourcegraph v6.4+ and v6.3.416, Claude 3.7 Extended Thinking - as well as Claude 4 base and extended thinking variants - are available in Sourcegraph when using Claude through either Google Vertex or AWS Bedrock.

See Model Configuration: Reasoning models for more information.

Autocomplete

Cody uses a set of models for autocomplete which are suited for the low latency use case.

ProviderModelFreeProEnterprise
Fireworks.aiDeepSeek-Coder-V2
Anthropicclaude Instant--
The default autocomplete model for Cody Free, Pro and Enterprise users is DeepSeek-Coder-V2.
The DeepSeek model used by Sourcegraph is hosted by Fireworks.ai, and is hosted as a single-tenant service in a US-based data center. For more information see our Cody FAQ.

Smart Apply

ProviderModelFreeProEnterprise
Fireworks.aiQwen 2.5 Coder

Fireworks.ai is the default model for cody-gateway, but if you wish to switch to Claude models, Site admins can do it following these steps-

Go to "Site admin"
Click on the "Feature flags"
Search for cody-smart-apply-instant-mode-enabled feature flag
Turn off/delete the "cody-smart-apply-instant-mode-enabled" feature flag
Previous
Proxy Setup