Skip to main content

πŸ”Œ Connect a Provider

Connect Open WebUI to any model provider and start chatting in minutes.

Open WebUI supports multiple connection protocols, including Ollama, OpenAI-compatible APIs, and Open Responses. Any cloud API or local server that speaks one of these protocols works out of the box. Just add a URL and API key, and your models appear in the dropdown.


How It Works​

β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”         β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”         β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚ β”‚ HTTP β”‚ β”‚ Inferenceβ”‚ β”‚
β”‚ Open WebUI │────────▢│ Provider API │────────▢ β”‚ Model β”‚
β”‚ (frontend) │◀────────│ (cloud/local) │◀──────── β”‚ (LLM/VLM) β”‚
β”‚ β”‚ Stream β”‚ β”‚ Tokens β”‚ β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜ β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
  1. You type a message in Open WebUI
  2. Open WebUI sends it to your provider's API endpoint
  3. The provider runs inference on the selected model
  4. Tokens stream back to Open WebUI in real time
  5. You see the response in the chat interface
tip

Adding a provider is as simple as entering a URL and API key in Admin Settings β†’ Connections. Open WebUI auto-detects available models from most providers.


Cloud Providers​

Hosted APIs that require an account and API key. No hardware needed.

ProviderModelsGuide
OllamaLlama, Mistral, Gemma, Phi, and thousands more (local)Starting with Ollama β†’
OpenAIGPT-4o, GPT-4.1, o3, o4-miniStarting with OpenAI β†’
AnthropicClaude Opus, Sonnet, HaikuStarting with Anthropic β†’
OpenAI-CompatibleGoogle Gemini, DeepSeek, Mistral, Groq, OpenRouter, Amazon Bedrock, Azure, and moreOpenAI-Compatible Providers β†’

Local Servers​

Run models on your own hardware. No API keys, no cloud dependency.

ServerDescriptionGuide
llama.cppEfficient GGUF model inference with OpenAI-compatible APIStarting with llama.cpp β†’
vLLMHigh-throughput inference engine for production workloadsStarting with vLLM β†’

More local servers (LM Studio, LocalAI, Docker Model Runner, Lemonade) are covered in the OpenAI-Compatible Providers guide.


Other Connection Methods​

FeatureDescriptionGuide
Open ResponsesConnect providers using the Open Responses specificationStarting with Open Responses β†’
FunctionsExtend Open WebUI with custom pipe functions for any backendStarting with Functions β†’

Looking for Agents?​

If you want to connect an autonomous AI agent (with terminal access, file operations, web search, and more) instead of a plain model provider, see Connect an Agent.