Skip to main content

Example of connecting to an inference service from Open WebUI

Last update:

Purpose of connection

Connect to the inference service from Open WebUI via a dedicated endpoint to interact with the model via a chatbot.

What you need to connect

  • inference-service;

  • Open WebUI Virtual Machine from AI Marketplace with configuration:

    • without a graphics processing unit (GPU);
    • on a private subnet with a single public IP address for access from the Internet.

Connection result

Open WebUI is connected to the inference service via a dedicated endpoint, allowing interaction with the model via a chatbot.

Connection steps

  1. Create an administrator account in Open WebUI.
  2. Start Open WebUI.
  3. Connect the endpoint in Open WebUI.
  4. Send a test request to the model.

1. Create an administrator account

To work in the Open WebUI, first create an administrator account.

  1. In the dashboard, from the top menu, click Products and select AI Marketplace.

  2. In the Open WebUI Virtual Machine card, click Go to GUI.

  3. In the Open WebUI, click Let's Get Started.

  4. Enter a user name.

  5. Enter email.

  6. Enter your password.

  7. Click Create an Administrator account.

2. Start Open WebUI

  1. In the dashboard, from the top menu, click Products and select AI Marketplace.

  2. In the Open WebUI Virtual Machine card, click Go to GUI.

  3. Enter the administrator's e-mail address.

  4. Enter the administrator's password.

  5. Click Sign In.

3. Connect the endpoint in Open WebUI

  1. In the Open WebUI web interface, click the profile icon and select Administrator Panel.

  2. Open the Settings tab.

  3. Go to the Connections section.

  4. Turn on the Direct Connections toggle switch.

  5. Click Save.

  6. On the main page of the Open WebUI web interface, click on the profile icon and select Settings.

  7. Go to the Connections section.

  8. In the Manage Direct Connections row, click .

  9. In the URL field, type endpoint:

    https://<inference_service_uuid>.wc.<pool>.inference.selcloud.ru/v1

    Specify:

    • <inference_service_uuuid> - The UUID of the inference service. The UUID can be copied from the control panel: in the top menu, click ProductsInference services → in the menu inference service menu select Copy UUID;

    • <pool> - pool where the inference service is created, e.g. ru-7.

  10. Enter the API key in the Bearer field. The API key can be copied in the control panel: in the top menu, click ProductsInference services → inference services page → API keys tab → in the API key row, click and then .

  11. In the Model IDs field, enter the model name and click . You can view the model name in the control panel: in the top menu, click ProductsInference Services → inference services page → Services tab → Model line.

  12. Click Save.

4. Send a test request to the model

  1. In the Open WebUI, click Select Model.

  2. Select the model you added in step 8 when connecting the endpoint in Open WebUI.

  3. On the main page of the Open WebUI web interface, click on the profile icon and select Settings.

  4. Go to the General section.

  5. In the Advanced Parameters block, set the max_tokens parameter to 5,000.

  6. Click Save.

  7. Enter a test query, such as:

    Привет!
  8. Click .

  9. Wait for the model to respond.