Skip to content

Custom OpenAI-Compatible Provider Endpoints (BYO Endpoint for Local & Remote LLMs) #9303

@tbitcs

Description

@tbitcs

Pre-submit Checks

Describe the solution you'd like?

Add support for configuring custom OpenAI-compatible API endpoints, allowing Warp to connect to local or self-hosted LLM providers.

This would enable users to specify:

The configured endpoint would be used for Warp AI features such as command suggestions, agent interactions, and inline completions.

The implementation should support standard OpenAI-compatible APIs, including chat completions and streaming responses, so that it works with tools like vLLM, Ollama (via compatibility layer), and other OpenAI-compatible servers.

From a user perspective, this would extend the existing “Bring Your Own Key” model into a “Bring Your Own Endpoint” workflow, where users can choose between built-in providers and a custom endpoint.

Is your feature request related to a problem? Please describe.

Yes.

Currently, Warp AI only supports a fixed set of providers (e.g. OpenAI, Anthropic, Google) via API keys. There is no way to connect Warp to local or self-hosted models.

This creates several limitations:

  • Users cannot run Warp AI in privacy-sensitive or offline environments
  • Teams with existing LLM infrastructure cannot integrate it
  • Developers cannot use open-source models running locally
  • All prompts must be sent to external providers, even when a local option is available

Many modern LLM tools (such as vLLM and Ollama) expose OpenAI-compatible APIs, but Warp cannot currently take advantage of this without native support for custom endpoints.

Additional context

This request consolidates several related feature requests:

The proposed approach is to extend Warp’s existing provider model to allow any OpenAI-compatible endpoint, rather than introducing provider-specific integrations for each tool.

This would allow immediate compatibility with:

  • vLLM (OpenAI-compatible server)
  • Ollama (local runtime with compatibility layer)
  • LiteLLM and similar proxy systems

Example configuration:

{
"provider": "custom",
"base_url": "http://localhost:8000/v1",
"model": "meta-llama/Llama-3-8b-instruct",
"api_key": "optional"
}

This is a provider-level feature and is expected to be cross-platform across all Warp-supported operating systems.

Operating system (OS)

Windows

How important is this feature to you?

5 (Can't work without it!)

Warp Internal (ignore) - linear-label:39cc6478-1249-4ee7-950b-c428edfeecd1

None

Metadata

Metadata

Assignees

No one assigned

    Labels

    area:agentAgent workflows, conversations, prompts, cloud mode, and AI-specific UI.area:settings-keybindingsSettings UI, preferences, keybindings, and keyboard-shortcut management.duplicateThis issue or pull request already exists.enhancementNew feature or request.repro:highThe report includes enough evidence that the issue appears highly reproducible.triagedIssue has received an initial automated triage pass.

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions