Supported LLMs
Chat and Prompts
Cody supports a variety of cutting-edge large language models for use in chat and prompts, allowing you to select the best model for your use case.
Provider | Model | Free | Pro | Enterprise | ||||
---|---|---|---|---|---|---|---|---|
OpenAI | GPT-4 Turbo | - | ✅ | ✅ | ||||
OpenAI | GPT-4o | - | ✅ | ✅ | ||||
OpenAI | GPT-4o-mini | ✅ | ✅ | ✅ | ||||
OpenAI | o3-mini-medium (experimental) | ✅ | ✅ | ✅ | ||||
OpenAI | o3-mini-high (experimental) | - | - | ✅ | ||||
OpenAI | o1 | - | ✅ | ✅ | ||||
Anthropic | Claude 3.5 Haiku | ✅ | ✅ | ✅ | ||||
Anthropic | Claude 3.5 Sonnet | ✅ | ✅ | ✅ | ||||
Anthropic | Claude 3.7 Sonnet | - | ✅ | ✅ | ||||
Gemini 1.5 Pro | ✅ | ✅ | ✅ (beta) | |||||
Gemini 2.0 Flash | ✅ | ✅ | ✅ | |||||
Gemini 2.0 Flash-Lite Preview (experimental) | ✅ | ✅ | ✅ |
Claude 3.7 Sonnet
Claude 3.7 has two variants — Claude 3.7 Sonnet and Claude 3.7 Extended Thinking — to support deep reasoning and fast, responsive edit workflows. This means you can use Claude 3.7 in different contexts depending on whether long-form reasoning is required or for tasks where speed and performance are a priority.
Claude 3.7 Extended Thinking is the recommended default chat model for Cloud customers. Self-hosted customers are encouraged to follow this recommendation, as Claude 3.7 outperforms 3.5 in most scenarios.
Claude 3.7 for GCP
In addition, Sourcegraph Enterprise customers using GCP Vertex (Google Cloud Platform) for Claude models can use both these variants of Claude 3.7 to optimize extended reasoning and deeper understanding. Customers using AWS Bedrock do not have the Claude 3.7 Extended Thinking variant.
Autocomplete
Cody uses a set of models for autocomplete which are suited for the low latency use case.
Provider | Model | Free | Pro | Enterprise | ||||
---|---|---|---|---|---|---|---|---|
Fireworks.ai | DeepSeek-Coder-V2 | ✅ | ✅ | ✅ | ||||
Fireworks.ai | StarCoder | - | - | ✅ | ||||
Anthropic | claude Instant | - | - | ✅ | ||||