edit or apply models are specified, respectively.
Recommended Chat models
Best overall experience
For the best overall Chat experience, you will want to use a 400B+ parameter model or one of the frontier models.Claude Opus 4.6 and Claude Sonnet 4 from Anthropic
Our current top recommendations are Claude Opus 4.6 and Claude Sonnet 4 from Anthropic.- Hub
- YAML
View the Claude Opus 4.6 model block or Claude Sonnet 4 model block on the hub.
Gemma from Google DeepMind
If you prefer to use an open-weight model, then the Gemma family of Models from Google DeepMind is a good choice. You will need to decide if you use it through a SaaS model provider, e.g. Together, or self-host it, e.g. Ollama.- Hub
- YAML
- Ollama
- Together
Add the Ollama Gemma 3 27B block from the hub
GPT-5.1 from OpenAI
If you prefer to use a model from OpenAI, then we recommend GPT-5.1.- Hub
- YAML
Add the OpenAI GPT-5.1 block from the hub
Grok-4 from xAI
If you prefer to use a model from xAI, then we recommend Grok-4.- Hub
- YAML
Add the xAI Grok-4.1 block from the hub
Gemini 3.1 Pro from Google
If you prefer to use a model from Google, then we recommend Gemini 3.1 Pro.- Hub
- YAML
Add the Gemini 3.1 Pro block from the hub
Local, Offline Experience
For the best local, offline Chat experience, you will want to use a model that is large but fast enough on your machine.Qwen 3 8B
If your local machine can run an 8B parameter model, then we recommend running Qwen 3 8B on your machine (e.g. using Ollama or LM Studio).- Hub
- YAML
- Ollama
Add the Ollama Qwen 3 8B block from the hub
Qwen 3 Coder 30B
If your local machine can run a larger model, then Qwen 3 Coder is an excellent code-specialized option (e.g. using Ollama or LM Studio). The 30B-A3B variant uses mixture-of-experts and runs efficiently despite its size.- YAML
- Ollama
- LM Studio
config.yaml