Models
Zed’s plans offer hosted versions of major LLMs, generally with higher rate limits than using your API keys. We’re working hard to expand the models supported by Zed’s subscription offerings, so please check back often.
Model | Provider | Token Type | Provider Price per 1M tokens | Zed Price per 1M tokens |
---|---|---|---|---|
Claude Opus 4.1 | Anthropic | Input | $15.00 | $16.50 |
Anthropic | Output | $75.00 | $82.50 | |
Anthropic | Input - Cache Write | $18.75 | $20.625 | |
Anthropic | Input - Cache Read | $1.50 | $1.65 | |
Claude Sonnet 4.5 | Anthropic | Input | $3.00 | $3.30 |
Anthropic | Output | $15.00 | $16.50 | |
Anthropic | Input - Cache Write | $3.75 | $4.125 | |
Anthropic | Input - Cache Read | $0.30 | $0.33 | |
Claude Sonnet 4 | Anthropic | Input | $3.00 | $3.30 |
Anthropic | Output | $15.00 | $16.50 | |
Anthropic | Input - Cache Write | $3.75 | $4.125 | |
Anthropic | Input - Cache Read | $0.30 | $0.33 | |
Claude Sonnet 3.7 | Anthropic | Input | $3.00 | $3.30 |
Anthropic | Output | $15.00 | $16.50 | |
Anthropic | Input - Cache Write | $3.75 | $4.125 | |
Anthropic | Input - Cache Read | $0.30 | $0.33 | |
GPT-5 | OpenAI | Input | $1.25 | $1.375 |
OpenAI | Output | $10.00 | $11.00 | |
OpenAI | Cached Input | $0.125 | $0.1375 | |
GPT-5 mini | OpenAI | Input | $0.25 | $0.275 |
OpenAI | Output | $2.00 | $2.20 | |
OpenAI | Cached Input | $0.025 | $0.0275 | |
GPT-5 nano | OpenAI | Input | $0.05 | $0.055 |
OpenAI | Output | $0.40 | $0.44 | |
OpenAI | Cached Input | $0.005 | $0.0055 | |
Gemini 2.5 Pro | Input | $1.25 | $1.375 | |
Output | $10.00 | $11.00 | ||
Gemini 2.5 Flash | Input | $0.30 | $0.33 | |
Output | $2.50 | $2.75 |
Usage
Any usage of a Zed-hosted model will be billed at the Zed Price (rightmost column above). See Plans and Usage for details on Zed's plans and limits for use of hosted models.
We encourage you to think through what model is best for your needs before leaving the Agent Panel to work. All LLMs can "spiral" and occasionally enter unending loops that require user intervention.
Context Windows
A context window is the maximum span of text and code an LLM can consider at once, including both the input prompt and output generated by the model.
Model | Provider | Zed-Hosted Context Window |
---|---|---|
Claude Opus 4.1 | Anthropic | 200k |
Claude Sonnet 4 | Anthropic | 200k |
Claude Sonnet 3.7 | Anthropic | 200k |
GPT-5 | OpenAI | 400k |
GPT-5 mini | OpenAI | 400k |
GPT-5 nano | OpenAI | 400k |
Gemini 2.5 Pro | 200k | |
Gemini 2.5 Flash | 200k |
We're planning on expanding supported context windows for hosted Sonnet 4 and Gemini 2.5 Pro/Flash in the near future. Stay tuned!
Each Agent thread and text thread in Zed maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.
For best results, it’s recommended you take a purpose-based approach to Agent thread management, starting a new thread for each unique task.
Tool Calls
Models can use tools to interface with your code, search the web, and perform other useful functions.