Supported LLMs
AI Assistant supports a variety of advanced cloud-based LLMs, as well as the option to use locally hosted models. This flexibility allows you to choose the most suitable model for your specific task. For example, you might want to use large models for complex codebase-related tasks, compact models for quick responses, or local models if you prefer to keep your data private.
Cloud-based models
The table below lists models available for selection in AI Assistant:
Model | Capabilities | Model context window |
|---|---|---|
| 200k | |
200k | ||
| 200k | |
200k | ||
200k | ||
200k | ||
200k | ||
1M | ||
1M | ||
1M | ||
1M | ||
| 400k[1] | |
| 400k[1] | |
| 400k[1] | |
1M | ||
1M | ||
1M | ||
128k | ||
| 200k | |
| 200k | |
200k | ||
| 200k |
- Supported models history
The following table lists AI models that have been available in AI Assistant, along with the plugin versions in which they were introduced or removed.
Model
Added in version
Removed in version
2025.2.x
–
2025.2.x
–
2025.2.x
–
251.26094.80.x
–
243.23654.270
–
243.23654.270
–
243.23654.270
–
251.23774.42.x
–
251.23774.42.28.x
–
251.26094.80.19
–
243.23654.270
–
2024.3
251.23774.42.28.x
2024.3
251.23774.42.28.x
2025.2
–
2025.2
–
2025.2
–
251.23774.42.28.x
–
251.23774.42.28.x
–
251.23774.42.28.x
–
2024.2
–
251.23774.42.28.x
–
251.23774.42.28.x
–
243.23654.270
–
243.23654.270
–
243.23654.270
251.23774.42.28.x
Local models
AI Assistant supports a selection of models available through Ollama, LM Studio, and other OpenAI-compatible endpoints like llama.cpp or LiteLLM. These models are optimized for local use, enabling powerful AI capabilities without the need for cloud access.
The default model context window for local models is set to 64 000 tokens. If needed, you can adjust this value in the settings.
For more information about setting up local models, refer to Use local models.