AI FAQs
Frequently asked questions about Warp's AI features, including supported models, privacy practices, request limits, billing, and usage guidelines.
General
What data is sent and/or stored when using Agents in Warp?
See our Privacy Page for more information on how we handle data used by Agents in Warp.
What happened to the old Warp AI chat panel?
Agent Mode has replaced the Warp AI chat panel. Agent Mode is more powerful in all of the chat panel's use cases. Not only can Agent Mode run commands for you, it can also gather context without you needing to copy and paste. To start a similar chat panel, click the AI button in the menu bar to start on a new AI pane.
Is my data used for model training?
No, Warp nor its providers (i.e. OpenAI, Anthropic, etc.) train on your data.
What model are you using for Agent Mode?
Warp supports a curated list of LLMs from providers like OpenAI, Anthropic, and Gemini. To view the full list of supported models and learn how to switch between them, visit the Model Choice page.
Can I use my own LLM API key?
Organizations on the Enterprise plan can enable a “Bring Your Own LLM” option to meet strict security or compliance requirements. Our team will work closely with you to support your preferred LLM provider. This feature is not currently available on other plans.
Billing
Every Warp plan includes a set number of AI requests per user per month. Please refer to pricing to compare plans.
AI Request limits apply to Agent Mode, Generate, and AI autofill in Workflows. When you have used up your allotted requests for the cycle, you will not be able to issue any more AI requests until the cycle renews.
For questions around what counts as a AI request, what counts as a token, and how often requests refresh, please refer to What counts as an AI request?and more on the Plans & Pricingpage.
Exceeding Agent Mode request limits
What is Lite?
Lite is a basic AI model included with the Turbo plan that serves two purposes:
Fallback model: If you reach your Turbo AI request limits, Warp automatically switches to Lite so you can keep using AI without interruption — at no additional cost.
Standalone option: You can also choose to use Lite before hitting your limits. In this case, usage will still count toward your monthly request limits, but once those limits are reached, Lite remains available with unlimited usage for Turbo plan users only.
Lite is a more token-efficient model than other premium models and supports core AI workflows. Learn more about Lite in the What is Lite? section of our Plans & Pricing documentation.
Common AI error messages
"Message token limit exceeded" error
This error means your input (plus attached context) exceeds the maximum context window of the model you're using. For example, GPT-4o has a context window limit of 123,904 tokens. If you exceed that, you may receive no output.
To fix this, try:
Starting a new conversation
Reducing the number of blocks or lines attached to your query
"Monthly request limit exceeded" error
Once you exceed your AI requests on the Turbo plan (see pricing for current limits), premium models will be disabled, and Warp will automatically switch you to Lite. This allows you to continue using AI features with a more token-efficient model until your quota resets at the start of your next billing cycle.
"Request failed with error: QuotaLimit" error
Once you exceed your AI token limits, all models will be disabled. Note that requests and tokens are calculated separately, and even though the plans may have a set number of requests, they also have a limited number of tokens.
"Request failed with error: ErrorStatus(403, "Your account has been blocked from using AI features")"
This message means your account has been blocked from using AI features, typically due to a violation of our Terms of Service or suspected abuse (e.g. attempting to bypass request or token limits).
To resolve or clarify this, please contact our team at [email protected] if you believe this was an error. We'll review your case and respond as soon as possible.
Gathering AI debugging ID
In cases where you have issues with the Agent, we may ask for the AI debugging ID to troubleshoot the specific conversation. To gather the debugging ID, RIGHT-CLICK
on the AI conversation block in question and select "Copy debugging ID", then paste that into the bug report that you submit so that our team can investigate the issue.
Last updated
Was this helpful?