Install drivers for GPU node groups in a Managed Kubernetes cluster
You can create Managed Kubernetes clusters on a cloud server with GPUs without pre-installed drivers. To install the driver yourself, use the NVIDIA® GPU Operator application.
-
Install the Helm package manager version 3.7.0 or higher.
-
Add the
nvidiarepository to Helm:helm repo add nvidia https://helm.ngc.nvidia.com/nvidia -
Update the
nvidiarepository in Helm:helm repo update -
Install NVIDIA GPU Operator and specify the correct version of the GPU driver:
helm install \--namespace gpu-operator \--create-namespace \--set driver.version=<driver_version> \gpu-operator nvidia/gpu-operatorSpecify
<driver_version>is the NVIDIA® driver version. You can look in the NVIDIA GPU Driver row in the GPU Operator Component Matrix table of the NVIDIA® documentation. -
To verify that NVIDIA GPU Operator and the GPU driver are installed correctly, run a GPU application. For example, the CUDA VectorAdd vector addition application:
cat << EOF | kubectl create -f -apiVersion: v1kind: Podmetadata:name: cuda-vectoraddspec:restartPolicy: OnFailurecontainers:- name: cuda-vectoraddimage: "nvidia/samples:vectoradd-cuda11.2.1"resources:limits:nvidia.com/gpu: 1EOF -
Check that the CUDA VectorAdd application has
completedsuccessfully - the status of the feed should beCompleted:kubectl get podsThe
cuda-vectoraddfeed will have a status ofCompletedin the response:NAME READY STATUS RESTARTS AGEcuda-vectoradd 0/1 Completed 0 51s