Skip to main content
Create a cloud server with GPU
Last update:

Create a cloud server with GPU

GPUs (graphics processing units) can be added to a cloud server — when creating a server or to an existing server.

GPUs are used as dedicated PCI devices inside the cloud server.

GPUs are available:

GPU configurations can only be used with the server's network boot disk.

Create a server with GPU

Use the Create cloud server instruction.

Select:

  • source is a finished image of Ubuntu 22.04 LTS Machine Learning 64-bit. The image contains the drivers needed to work with graphics processors. If you choose another source, you will need to install the drivers yourself;
  • configuration — with vCPUs from 2 cores: fixed GPU Line configuration or arbitrary.

Add a GPU to an existing cloud server

If the cloud server has an arbitrary configuration and a network boot disk, GPUs can be added to it.

  1. In Control Panel, go to Cloud PlatformServers.
  2. Open the server page → Configuration tab.
  3. In the Configuration box, select an arbitrary configuration.
  4. Click Add GPU.
  5. Specify the number of GPUs.
  6. Click Save and Reload.

Available GPUs

NVIDIA® A100

NVIDIA® A100 NVLink
NVIDIA® Tesla T4NVIDIA® A30NVIDIA® A2
(updated
NVIDIA® Tesla T4 analog)
NVIDIA® GTX 1080NVIDIA® A2000
(RTX 3060 analog)
NVIDIA® A5000
(RTX 3080 analog)
Memory40GB
HBM2
16GB
GDDR6
24GB
HBM2
16GB
GDDR6
8GB
GDDR5X
6GB
GDDR6
24GB
GDDR6
CUDA cores6192256038041280256033288192
Tensor cores43232022440104256

You can view the current list of GPUs in control panel: under Cloud PlatformServers → click Create Server.

You can see the availability of GPUs in regions in the [GPUs for cloud servers] availability matrix(/control-panel-actions/availability-matrix.mdx#gpu-for-cloud-and-kubernetes).

NVIDIA® A100

Has maximum performance for AI, HPC, and data processing. Suitable for deep learning, scientific research and data analytics.

Ampere® based, up to 2 TB/s throughput. Refer to NVIDIA® documentation for detailed specifications.

Fixed GPU Line configurations are available from 1 to 8 GPUs × 40 GB, with vCPUs from 6 to 48, RAM from 87 to 700 GB.

Random configurations are 1 to 8 GPUs × 40 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.

You can combine two NVIDIA® A100 GPUs using NVLink technology.

NVLink accelerates the communication speed of GPU interconnect compared to the PCIe interface. NVLink-coupled GPUs allow more memory to be utilized and improve server performance for complex computations, such as training large language ML models.

NVLink works with NVIDIA® A100 — Ampere®-based GPUs with up to 2 TB/s of bandwidth. For detailed specifications of the NVIDIA® A100 and a description of NVLink technology, refer to the NVIDIA® documentation.

In fixed GPU Line configurations, 2 GPUs × 40 GB are available, with 24 vCPUs, 128 GB RAM. If configurations with more resources are needed, create a ticket.

NVIDIA® Tesla T4

Suitable for Machine Learning and Deep Learning, inference, graphics work and video rendering. Works with most AI frameworks and is compatible with all types of neural networks.

Based on Turing®, up to 300 GB/s throughput. Refer to NVIDIA® documentation for detailed specifications.

In fixed GPU Line configurations, 1 to 4 GPUs × 16 GB are available, with vCPUs from 4 to 24, RAM from 32 to 320 GB.

Random configurations are 1 to 4 GPUs × 16 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.

NVIDIA® A30

Suitable for AI-inference, HPC, language processing, conversational artificial intelligence, recommender systems.

Ampere® based, up to 933 GB/s bandwidth. Refer to NVIDIA® documentation for detailed specifications.

Fixed GPU Line configurations are available from 1 to 2 GPUs, with vCPUs from 16 to 48, RAM from 64 to 320 GB.

In random configurations, 1 to 2 GPUs, with vCPUs from 2 to 32, RAM from 512MB to 256GB.

NVIDIA® A2

Entry-level GPU. Suitable for simple inference, video and graphics, Edge AI (edge computing), Edge video, mobile cloud gamification.

Ampere® based, up to 200 GB/s bandwidth. Refer to NVIDIA® documentation for detailed specifications.

In fixed GPU Line configurations, 1 to 4 GPUs × 16 GB are available, with vCPUs from 12 to 48, RAM from 32 to 320 GB.

Random configurations are 1 to 4 GPUs × 16 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.

NVIDIA® GTX 1080

Performance and power efficient GPU. The solution is implemented with FinFET technology and GDDR5X memory. Dynamic load balancing helps to separate tasks so that resources don't sit idle waiting. Featuring maximum performance for information display, VR, ultra-high resolution settings and data processing.

Pascal® based, up to 320.3 Gb/s throughput. Refer to NVIDIA® documentation for detailed specifications.

In fixed GPU Line configurations, 1 to 8 GPUs × 8 GB are available, with vCPUs from 8 to 28, RAM from 24 to 96 GB.

Random configurations are 1 to 8 GPUs × 8 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.

NVIDIA® A2000

Power-efficient GPU for compact workstations. Suitable for AI, graphics and video rendering.

Ampere® based, up to 288 GB/s bandwidth. Refer to NVIDIA® documentation for detailed specifications.

In fixed GPU Line configurations, 1 to 4 GPUs × 6 GB are available, with vCPUs from 6 to 24, RAM from 16 to 320 GB.

Random configurations are 1 to 4 GPUs × 6 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.

NVIDIA® A5000

A versatile GPU, suitable for any task within its performance limits.

Ampere® based, up to 768 GB/s bandwidth. Refer to NVIDIA® documentation for detailed specifications.

In fixed GPU Line configurations, 1 to 2 GPUs × 24 GB are available, with vCPUs from 8 to 48, RAM from 32 to 320 GB.

Random configurations are 1 to 2 GPUs × 24 GB, with vCPUs from 2 to 32, RAM from 512 MB to 256 GB.