Before you start
Meet the following prerequisites, depending on the preferred interface:- Web console
- CLI
Create a key pair for SSH access to the VM and save the key pair to the default location:
Create the VM
- Web console
- CLI
- Go to the web console, click Create resource and then select Virtual machine.
-
On the VM creation page that opens, set the following parameters:
- Platform: NVIDIA® H100 NVLink with Intel Sapphire Rapids.
- Preset: 1 GPU - 16 CPUs - 200 GiB RAM.
- Boot disk image: Ubuntu 22.04 LTS for NVIDIA® GPUs (CUDA® 12). For details about boot disk images, see Boot disk images for Compute virtual machines.
- Boot disk size: 300 GiB SSD.
- Network: Select the Public IP address: Auto assign static IP option.
-
Username and SSH key: Select the public key that you created earlier.
In this field, do not use the
rootoradminusernames. They are reserved for internal needs and are not allowed to connect to a VM by SSH.
- Click Create VM.
Connect to the VM
-
Get the public IP address of the VM:
- Web console
- CLI
- Open the VM page.
- In the Network block, copy the Public IPv4 value.
-
Connect to the VM:
Specify the received public IP address and the username that you set during the VM creation.
Create a virtual environment and install the necessary packages
To work with Open WebUI, you need a dedicated virtual environment. It enables you to set up and run the OpenWebUI server in isolation from other software on the VM. To create a virtual environment, use Miniconda. To configure the environment:-
Download and install the latest Miniconda version:
-
Initialize Miniconda:
On initialization, Miniconda activates its
baseenvironment. -
Create an
OpenWebUIenvironment with Python 3.11:This command creates and activates a new environment. -
Install Ollama, which provides access to the model:
-
Install Open WebUI:
Start the Open WebUI server
-
Start the server:
-
Open the Open WebUI interface in the browser. To do this, enter the
http://<public_ip_address>:8080address in the search bar. - In the Open WebUI interface, create an account to work with LLMs locally within the VM. For details on working with Open WebUI, see their documentation.
open-webui serve.
Download the Qwen/Qwen2.5-72B-Instruct model
- In Open WebUI, click Select a model.
- Paste
qwen2.5:72binto the search bar. - Click Pull “qwen2.5:72b” from Ollama and wait for the download to finish.
- Click Select a model again and then choose Qwen/Qwen2.5-72B-Instruct.
Make Open WebUI start automatically
With the current configuration, you need to manually start the Open WebUI server every time you connect to your VM. Alternatively, you can configure the server to start up whenever the VM starts. To do this:-
Create a
systemdservice file for Open WebUI and open the file in an editor: -
Paste the following contents into the file and save it. Specify the username that you set during the VM creation:
-
To make the new service file recognizable, reload
systemd: -
To start automatically and immediately, enable the
systemdservice: -
Verify that the service is running:
http://ip_address:8080 and work with the Qwen/Qwen2.5-72B-Instruct model.