Nvidia gpu operator

icon

Enable GPU acceleration for compute workloads on NVIDIA GPUs. https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/latest/gpu-sharing.html


plugins/nvidia-gpu-operator/terra.yaml
resource_id: nvidia-gpu-operator
name: NVIDIA GPU Operator
icon: https://www.nvidia.com/content/nvidiaGDC/us/en_US/about-nvidia/legal-info/logo-brand-usage/_jcr_content/root/responsivegrid/nv_container_392921705/nv_container/nv_image.coreimg.100.410.png/1703060329053/nvidia-logo-vert.png
description: |
  Enable GPU acceleration for compute workloads on NVIDIA GPUs. https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/latest/gpu-sharing.html
category: Hardware
tags:
  - gpu
  - nvidia
  - time-slicing
  - vm-passthrough
  - operator
  - cluster-level
editable: true
fields:
  - name: install_crds
    description: |
      Install the NVIDIA GPU Operator's Custom Resource Definitions (CRDs). 
      If this is the first or only install of the NVIDIA GPU Operator in 
      the cluster, you should set this to true to install the necessary CRDs. 
      If you have already installed the NVIDIA GPU Operator with CRD installation 
      enabled, you can set this to false for subsequent installations to avoid 
      trying to reinstall the CRDs.
    required: false
    default: false
    type: boolean
  - name: open_kernel_modules
    description: |
      Whether to use open kernel modules for the NVIDIA driver. This is required 
      if you want to use the open kernel modules for the NVIDIA driver. Not needed 
      if you are using the device plugin to manage the driver installation.
    required: false
    default: false
    type: boolean
  - name: k3s
    description: |
      When installing in a k3s cluster, the NVIDIA GPU Operator needs to be configured 
      to point to the correct containerd socket. The path is: /run/k3s/containerd/containerd.sock. 
      Setting this to true will configure the NVIDIA GPU Operator to use this socket path for 
      containerd, which is necessary for it to function properly in a k3s environment.
    required: false
    default: true
    type: boolean
  - name: slice_count
    description: |
      Number of time-slices to create for each GPU. This is required for time-slicing 
      configuration and determines how many slices each GPU will be divided into. The 
      default is 4, which means that each GPU will be divided into 4 time-slices. You 
      can adjust this number based on your workload requirements and the capabilities 
      of your GPUs.
    required: true
    default: 4
    type: int
  - name: version
    description: Version of the GPU Operator to install.
    required: true
    type: select
    options:
      - v25.10.1
      - v25.3.4