Example of connecting to an inference service from Open WebUI
Purpose of connection
Connect to the inference service from Open WebUI via a dedicated endpoint to interact with the model via a chatbot.
What you need to connect
-
inference-service;
-
Open WebUI Virtual Machine from AI Marketplace with configuration:
- without a graphics processing unit (GPU);
- on a private subnet with a single public IP address for access from the Internet.
Connection result
Open WebUI is connected to the inference service via a dedicated endpoint, allowing interaction with the model via a chatbot.
Connection steps
- Create an administrator account in Open WebUI.
- Start Open WebUI.
- Connect the endpoint in Open WebUI.
- Send a test request to the model.
1. Create an administrator account
To work in the Open WebUI, first create an administrator account.
-
In the dashboard, from the top menu, click Products and select AI Marketplace.
-
In the Open WebUI Virtual Machine card, click Go to GUI.
-
In the Open WebUI, click Let's Get Started.
-
Enter a user name.
-
Enter email.
-
Enter your password.
-
Click Create an Administrator account.
2. Start Open WebUI
-
In the dashboard, from the top menu, click Products and select AI Marketplace.
-
In the Open WebUI Virtual Machine card, click Go to GUI.
-
Enter the administrator's e-mail address.
-
Enter the administrator's password.
-
Click Sign In.
3. Connect the endpoint in Open WebUI
-
In the Open WebUI web interface, click the profile icon and select Administrator Panel.
-
Open the Settings tab.
-
Go to the Connections section.
-
Turn on the Direct Connections toggle switch.
-
Click Save.
-
On the main page of the Open WebUI web interface, click on the profile icon and select Settings.
-
Go to the Connections section.
-
In the Manage Direct Connections row, click .
-
In the URL field, type endpoint:
<endpoint>/v1Specify
<endpoint>- the endpoint of the inference service.The endpoint can be copied in the control panel: in the top menu, click Products → Inference Services → inference service page → Quick Start tab → in the Endpoint block, click . -
In the Bearer field, enter the API key.The API key can be copied in the control panel: in the top menu, click Products → Inference Services → inference services page → API keys tab → in the API key row, click , and then click .
-
In the Model IDs field, enter the model name and click .You can view the model name in the control panel: in the top menu, click Products → Inference Services → inference services page → Services tab → Model line.
-
Click Save.
4. Send a test request to the model
-
In the Open WebUI, click Select Model.
-
Select the model you added in step 8 when connecting the endpoint in Open WebUI.
-
On the main page of the Open WebUI web interface, click on the profile icon and select Settings.
-
Go to the General section.
-
In the Advanced Parameters block, set the
max_tokensparameter to5,000. -
Click Save.
-
Enter a test query, such as:
Привет! -
Click .
-
Wait for the model to respond.