Skip to content

Agents > Capabilities

Agent model choice

Open in ChatGPT ↗
Ask ChatGPT about this page
Open in Claude ↗
Ask Claude about this page
Copied!

Choose from a curated set of top LLMs for Warp's Agents (or let Warp auto-select the best model).

Warp lets you choose from a curated set of Large Language Models (LLMs) to power your Agentic Development Environment.

Warp supports the following models.

The model_id values shown below can be used when configuring models via the Oz Platform or CLI.

Modelmodel_idDescription
Auto (Responsive)autoSelects the highest-quality, fastest available model. May consume credits more quickly.
Auto (Cost-efficient)auto-efficientOptimizes for lower credit consumption while maintaining strong output quality.
Auto (Genius)auto-geniusAdapts to task complexity and selects Warp’s most capable model when it’s worth it. Best for deep debugging, architecture decisions, and /plan sessions.
Auto (Open-weights)auto-openRoutes between the best open-source models available in Warp. Optimizes for low cost and fast speed using open-weights models.

All Auto models perform well across all agent workflows and are ideal if you prefer Warp to manage model selection dynamically.

Modelmodel_idReasoning Level
GPT-5.5gpt-5-5-lowLow
GPT-5.5gpt-5-5-mediumMedium
GPT-5.5gpt-5-5-highHigh
GPT-5.5gpt-5-5-xhighExtra High
GPT-5.4gpt-5-4-lowLow
GPT-5.4gpt-5-4-mediumMedium
GPT-5.4gpt-5-4-highHigh
GPT-5.4gpt-5-4-xhighExtra High
GPT-5.3 Codexgpt-5-3-codex-lowLow
GPT-5.3 Codexgpt-5-3-codex-mediumMedium
GPT-5.3 Codexgpt-5-3-codex-highHigh
GPT-5.3 Codexgpt-5-3-codex-xhighExtra High
GPT-5.2 Codexgpt-5-2-codex-lowLow
GPT-5.2 Codexgpt-5-2-codex-mediumMedium
GPT-5.2 Codexgpt-5-2-codex-highHigh
GPT-5.2 Codexgpt-5-2-codex-xhighExtra High
GPT-5.2gpt-5-2-lowLow
GPT-5.2gpt-5-2-mediumMedium
GPT-5.2gpt-5-2-highHigh
GPT-5.2gpt-5-2-xhighExtra High
Modelmodel_idVariant
Claude Opus 4.7claude-4-7-opus-xhighDefault effort
Claude Opus 4.7claude-4-7-opus-highHigh effort
Claude Opus 4.7claude-4-7-opus-maxMax effort
Claude Opus 4.6claude-4-6-opus-highDefault effort
Claude Opus 4.6claude-4-6-opus-maxMax effort
Claude Sonnet 4.6claude-4-6-sonnet-highDefault effort
Claude Sonnet 4.6claude-4-6-sonnet-maxMax effort
Claude Opus 4.5claude-4-5-opusThinking off
Claude Opus 4.5claude-4-5-opus-thinkingThinking on
Claude Sonnet 4.5claude-4-5-sonnetThinking off
Claude Sonnet 4.5claude-4-5-sonnet-thinkingThinking on
Claude Haiku 4.5claude-4-5-haiku
Modelmodel_id
Gemini 3.1 Progemini-3.1-pro

Warp also supports leading open source models hosted via Fireworks AI, so you can run them from inside Warp without setting up your own inference infrastructure.

Modelmodel_id
GLM 5glm-5-fireworks
GLM 5.1glm-5.1-fireworks
Kimi K2.5kimi-k25-fireworks
Kimi K2.6kimi-k26-fireworks
Minimax 2.7minimax-2.7-fireworks
Qwen 3.6 Plusqwen-3.6-plus-fireworks

You can use the model picker in your prompt input to quickly switch between models. The currently active model appears directly in the input editor.

Model selector dropdown showing available models with Intelligence, Speed, and Cost benchmarks

Model selector in Warp’s input.

To change models, click the displayed model name (for example, Claude Sonnet 4.5) to open a dropdown with all supported options. Your selection will automatically persist for future prompts.

Warp uses a model fallback system to ensure uninterrupted service if your selected model becomes temporarily unavailable due to provider outages or capacity issues.

How it works:

  • If your selected model isn’t available, Warp automatically uses a fallback model from a predefined chain to continue your conversation without errors.
  • As soon as your originally selected model becomes available again, Warp automatically switches back to it.
  • The fallback model is selected to provide comparable quality and capabilities to your original choice.

You can configure the base model for each Agent Profile, alongside the Agent’s autonomy, tool access, and other permissions. The base model is also used for Planning.

Edit your default profile or any other profile directly in Settings > Agents > Profiles.

Warp integrates with multiple Large Language Model (LLM) providers to power its AI-driven features.

These providers include, but are not limited to:

  • OpenAI
  • Anthropic
  • Google
  • xAI
  • Fireworks AI

Warp has executed Zero Data Retention (ZDR) agreements with these providers. This means that, by default across all plans:

  • LLM providers commit not to train their models on any customer-generated data processed through Warp’s services.
  • LLM providers commit to delete inputs and outputs after generating the relevant output, within a fixed time period.

Warp enforces these commitments through both technical measures and contractual safeguards with the LLM providers.