Keep Your Open WebUI Healthy with Monitoring π©Ί
Monitoring your Open WebUI instance is crucial for ensuring it runs reliably, performs well, and allows you to quickly identify and resolve any issues. This guide outlines three levels of monitoring, from basic availability checks to in-depth model response testing.
Why Monitor?
- Ensure Uptime: Proactively detect outages and service disruptions.
- Performance Insights: Track response times and identify potential bottlenecks.
- Early Issue Detection: Catch problems before they impact users significantly.
- Peace of Mind: Gain confidence that your Open WebUI instance is running smoothly.
π¦ Levels of Monitoringβ
We'll cover three levels of monitoring, progressing from basic to more comprehensive:
- Basic Health Check: Verifies if the Open WebUI service is running and responding.
- Model Connectivity Check: Confirms that Open WebUI can connect to and list your configured models.
- Model Response Testing (Deep Health Check): Ensures that models can actually process requests and generate responses.
Level 1: Basic Health Check Endpoint β β
The simplest level of monitoring is checking the /health
endpoint. This endpoint is publicly accessible (no authentication required) and returns a 200 OK
status code when the Open WebUI service is running correctly.
How to Test:
You can use curl
or any HTTP client to check this endpoint:
# Basic health check - no authentication needed
curl https://your-open-webui-instance/health
Expected Output: A successful health check will return a 200 OK
HTTP status code. The content of the response body is usually not important for a basic health check.
Using Uptime Kuma for Basic Health Checks π»β
Uptime Kuma is a fantastic, open-source, and easy-to-use self-hosted uptime monitoring tool. It's highly recommended for monitoring Open WebUI.
Steps to Set Up in Uptime Kuma:
- Add a New Monitor: In your Uptime Kuma dashboard, click "Add New Monitor".
- Configure Monitor Settings:
- Monitor Type: Select "HTTP(s)".
- Name: Give your monitor a descriptive name, e.g., "Open WebUI Health Check".
- URL: Enter the health check endpoint URL:
http://your-open-webui-instance:8080/health
(Replaceyour-open-webui-instance:8080
with your actual Open WebUI address and port). - Monitoring Interval: Set the frequency of checks (e.g.,
60 seconds
for every minute). - Retry Count: Set the number of retries before considering the service down (e.g.,
3
retries).
What This Check Verifies:
- Web Server Availability: Ensures the web server (e.g., Nginx, Uvicorn) is responding to requests.
- Application Running: Confirms that the Open WebUI application itself is running and initialized.
- Basic Database Connectivity: Typically includes a basic check to ensure the application can connect to the database.
Level 2: Open WebUI Model Connectivity πβ
To go beyond basic availability, you can monitor the /api/models
endpoint. This endpoint requires authentication and verifies that Open WebUI can successfully communicate with your configured model providers (e.g., Ollama, OpenAI) and retrieve a list of available models.
Why Monitor Model Connectivity?
- Model Provider Issues: Detect problems with your model provider services (e.g., API outages, authentication failures).
- Configuration Errors: Identify misconfigurations in your model provider settings within Open WebUI.
- Ensure Model Availability: Confirm that the models you expect to be available are actually accessible to Open WebUI.
API Endpoint Details:
See the Open WebUI API documentation for full details about the /api/models
endpoint and its response structure.
How to Test with curl
(Authenticated):
You'll need an API key to access this endpoint. See the "Authentication Setup" section below for instructions on generating an API key.
# Authenticated model connectivity check
curl -H "Authorization: Bearer YOUR_API_KEY" https://your-open-webui-instance/api/models
(Replace YOUR_API_KEY
with your actual API key and your-open-webui-instance
with your Open WebUI address.)
Expected Output: A successful request will return a 200 OK
status code and a JSON response containing a list of models.
Authentication Setup for API Key πβ
Before you can monitor the /api/models
endpoint, you need to enable API keys in Open WebUI and generate one:
-
Enable API Keys (Admin Required):
- Log in to Open WebUI as an administrator.
- Go to Admin Settings (usually in the top right menu) > General.
- Find the "Enable API Key" setting and turn it ON.
- Click Save Changes.
-
Generate an API Key (User Settings):
- Go to your User Settings (usually by clicking on your profile icon in the top right).
- Navigate to the Account section.
- Click Generate New API Key.
- Give the API key a descriptive name (e.g., "Monitoring API Key").
- Copy the generated API key and store it securely. You'll need this for your monitoring setup.
(Optional but Recommended): For security best practices, consider creating a non-administrator user account specifically for monitoring and generate an API key for that user. This limits the potential impact if the monitoring API key is compromised.
If you don't see the API key generation option in your settings, contact your Open WebUI administrator to ensure API keys are enabled.
Using Uptime Kuma for Model Connectivity Monitoring π»β
-
Create a New Monitor in Uptime Kuma:
- Monitor Type: "HTTP(s) - JSON Query".
- Name: "Open WebUI Model Connectivity Check".
- URL:
http://your-open-webui-instance:8080/api/models
(Replace with your URL). - Method: "GET".
- Expected Status Code:
200
.
-
Configure JSON Query (Verify Model List):
- JSON Query:
$count(data[*])>0
- Explanation: This JSONata query checks if the
data
array in the API response (which contains the list of models) has a count greater than 0. In other words, it verifies that at least one model is returned.
- Explanation: This JSONata query checks if the
- Expected Value:
true
(The query should returntrue
if models are listed).
- JSON Query:
-
Add Authentication Headers:
- In the "Headers" section of the Uptime Kuma monitor configuration, click "Add Header".
- Header Name:
Authorization
- Header Value:
Bearer YOUR_API_KEY
(ReplaceYOUR_API_KEY
with the API key you generated).
-
Set Monitoring Interval: Recommended interval:
300 seconds
(5 minutes) or longer, as model lists don't typically change very frequently.
Alternative JSON Queries (Advanced):
You can use more specific JSONata queries to check for particular models or providers. Here are some examples:
- Check for at least one Ollama model:
$count(data[owned_by='ollama'])>0
- Check if a specific model exists (e.g., 'gpt-4o'):
$exists(data[id='gpt-4o'])
- Check if multiple specific models exist (e.g., 'gpt-4o' and 'gpt-4o-mini'):
$count(data[id in ['gpt-4o', 'gpt-4o-mini']]) = 2
You can test and refine your JSONata queries at jsonata.org using a sample API response to ensure they work as expected.
Level 3: Model Response Testing (Deep Health Check) π€β
For the most comprehensive monitoring, you can test if models are actually capable of processing requests and generating responses. This involves sending a simple chat completion request to the /api/chat/completions
endpoint.
Why Test Model Responses?
- End-to-End Verification: Confirms that the entire model pipeline is working, from API request to model response.
- Model Loading Issues: Detects problems with specific models failing to load or respond.
- Backend Processing Errors: Catches errors in the backend logic that might prevent models from generating completions.
How to Test with curl
(Authenticated POST Request):
This test requires an API key and sends a POST request with a simple message to the chat completions endpoint.
# Test model response - authenticated POST request
curl -X POST https://your-open-webui-instance/api/chat/completions \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"messages": [{"role": "user", "content": "Respond with the word HEALTHY"}],
"model": "llama3.1", # Replace with a model you expect to be available
"temperature": 0 # Set temperature to 0 for consistent responses
}'
(Replace YOUR_API_KEY
, your-open-webui-instance
, and llama3.1
with your actual values.)
Expected Output: A successful request will return a 200 OK
status code and a JSON response containing a chat completion. You can verify that the response includes the word "HEALTHY" (or a similar expected response based on your prompt).
Setting up Level 3 Monitoring in Uptime Kuma would involve configuring an HTTP(s) monitor with a POST request, JSON body, authentication headers, and potentially JSON query to validate the response content. This is a more advanced setup and can be customized based on your specific needs.
By implementing these monitoring levels, you can proactively ensure the health, reliability, and performance of your Open WebUI instance, providing a consistently positive experience for users.