-
Notifications
You must be signed in to change notification settings - Fork 1.8k
Custom OpenAI-Compatible Provider Endpoints (BYO Endpoint for Local & Remote LLMs) #9303
Copy link
Copy link
Open
Labels
area:agentAgent workflows, conversations, prompts, cloud mode, and AI-specific UI.Agent workflows, conversations, prompts, cloud mode, and AI-specific UI.area:settings-keybindingsSettings UI, preferences, keybindings, and keyboard-shortcut management.Settings UI, preferences, keybindings, and keyboard-shortcut management.duplicateThis issue or pull request already exists.This issue or pull request already exists.enhancementNew feature or request.New feature or request.repro:highThe report includes enough evidence that the issue appears highly reproducible.The report includes enough evidence that the issue appears highly reproducible.triagedIssue has received an initial automated triage pass.Issue has received an initial automated triage pass.
Metadata
Metadata
Assignees
Labels
area:agentAgent workflows, conversations, prompts, cloud mode, and AI-specific UI.Agent workflows, conversations, prompts, cloud mode, and AI-specific UI.area:settings-keybindingsSettings UI, preferences, keybindings, and keyboard-shortcut management.Settings UI, preferences, keybindings, and keyboard-shortcut management.duplicateThis issue or pull request already exists.This issue or pull request already exists.enhancementNew feature or request.New feature or request.repro:highThe report includes enough evidence that the issue appears highly reproducible.The report includes enough evidence that the issue appears highly reproducible.triagedIssue has received an initial automated triage pass.Issue has received an initial automated triage pass.
Pre-submit Checks
Describe the solution you'd like?
Add support for configuring custom OpenAI-compatible API endpoints, allowing Warp to connect to local or self-hosted LLM providers.
This would enable users to specify:
The configured endpoint would be used for Warp AI features such as command suggestions, agent interactions, and inline completions.
The implementation should support standard OpenAI-compatible APIs, including chat completions and streaming responses, so that it works with tools like vLLM, Ollama (via compatibility layer), and other OpenAI-compatible servers.
From a user perspective, this would extend the existing “Bring Your Own Key” model into a “Bring Your Own Endpoint” workflow, where users can choose between built-in providers and a custom endpoint.
Is your feature request related to a problem? Please describe.
Yes.
Currently, Warp AI only supports a fixed set of providers (e.g. OpenAI, Anthropic, Google) via API keys. There is no way to connect Warp to local or self-hosted models.
This creates several limitations:
Many modern LLM tools (such as vLLM and Ollama) expose OpenAI-compatible APIs, but Warp cannot currently take advantage of this without native support for custom endpoints.
Additional context
This request consolidates several related feature requests:
The proposed approach is to extend Warp’s existing provider model to allow any OpenAI-compatible endpoint, rather than introducing provider-specific integrations for each tool.
This would allow immediate compatibility with:
Example configuration:
{
"provider": "custom",
"base_url": "http://localhost:8000/v1",
"model": "meta-llama/Llama-3-8b-instruct",
"api_key": "optional"
}
This is a provider-level feature and is expected to be cross-platform across all Warp-supported operating systems.
Operating system (OS)
Windows
How important is this feature to you?
5 (Can't work without it!)
Warp Internal (ignore) - linear-label:39cc6478-1249-4ee7-950b-c428edfeecd1
None