AI FAQs
Frequently asked questions about Warp's AI features, including supported models, privacy practices, request limits, billing, and usage guidelines.
General
What data is sent and/or stored when using Agents in Warp?
See our Privacy Page for more information on how we handle data used by Agents in Warp
What happened to the old Warp AI chat panel?
Agent Mode has replaced the Warp AI chat panel. Agent Mode is more powerful in all of the chat panel's use cases. Not only can Agent Mode run commands for you, it can also gather context without you needing to copy and paste. To start a similar chat panel, click the AI button in the menu bar to start on a new AI pane.
Is my data used for model training?
No, Warp nor its providers (i.e. OpenAI, Anthropic, etc.) train on your data.
What model are you using for Agent Mode?
Warp supports a curated list of LLMs from providers like OpenAI, Anthropic, and Gemini. To view the full list of supported models and learn how to switch between them, visit the Model Choice page.
Can I use my own LLM API key?
Organizations on the Enterprise plan can enable a “Bring Your Own LLM” option to meet strict security or compliance requirements. Our team will work closely with you to support your preferred LLM provider. This feature is not currently available on other plans.
Billing
Every Warp plan includes a set number of AI requests per user per month. Please refer to pricing to compare plans.
AI Request limits apply to Agent Mode, Generate, and AI autofill in Workflows. When you have used up your allotted requests for the cycle, you will not be able to issue any more AI requests until the cycle renews.
For questions around what counts as a AI request, what counts as a token, and how often requests refresh, please refer to What counts as an AI request?and more on the Plans & Pricingpage.
Exceeding Agent Mode request limits
What is Lite?
Lite is a basic AI model included with the Turbo plan that serves two purposes:
Fallback model: If you reach your Turbo AI request limits, Warp automatically switches to Lite so you can keep using AI without interruption — at no additional cost.
Standalone option: You can also choose to use Lite before hitting your limits. In this case, usage will still count toward your monthly request limits, but once those limits are reached, Lite remains available with unlimited usage for Turbo plan users only.
Lite is a more token-efficient model than other premium models and supports core AI workflows. Learn more about Lite in the What is Lite? section of our Plans & Pricing documentation.
"Message token limit exceeded" error
This error means your input (plus attached context) exceeds the maximum context window of the model you're using. For example, GPT-4o has a context window limit of 123,904 tokens. If you exceed that, you may receive no output.
To fix this, try:
Starting a new conversation
Reducing the number of blocks or lines attached to your query
"Monthly request limit exceeded" error
Once you exceed your AI requests on the Turbo plan (see pricing for current limits), premium models will be disabled, and Warp will automatically switch you to Lite. This allows you to continue using AI features with a more token-efficient model until your quota resets at the start of your next billing cycle.
"Request failed with error: QuotaLimit" error
Once you exceed your AI token limits, all models will be disabled. Note that requests and tokens are calculated separately, and even though the plans may have a set number of requests, they also have a limited number of tokens.
If you have questions or need extended access, feel free to reach out to us at [email protected].
Last updated
Was this helpful?