Control Plane System Requirements

The NVIDIA Run:ai control plane is a Kubernetes-based application that centrally manages workloads, users, scheduling, and cluster integrations across multiple tenants.

In a multi-tenant deployment, the control plane is installed once, in a dedicated Kubernetes cluster and configured for multi-tenancy. This section outlines the hardware and software system requirements needed to deploy and operate the control plane in a multi-tenant environment.

Installer Machine

The machine running the installation script (typically the Kubernetes master) must have:

Hardware Requirements

The following hardware requirements are for the control plane system nodes. By default, all NVIDIA Run:ai control plane services run on all available nodes.

Architecture

x86 and ARM architectures are supported for Kubernetes.

NVIDIA Run:ai Control Plane - System Nodes

This configuration is the minimum requirement you need to install and use NVIDIA Run:ai control plane:

Component
Required Capacity

CPU

10 cores

Memory

12GB

Disk space

110GB

circle-info

Note

  • To designate nodes to NVIDIA Run:ai system services, follow the instructions as described in System nodes.

  • If you are using Grafana Mimir for monitoring, we recommend using the Microservices mode to properly size your environment. Refer to the following guide for capacity planning, Planning Grafana Mimir capacityarrow-up-right.

Software Requirements

The following software requirements must be fulfilled.

Operating System

  • Any Linux operating system supported by both Kubernetes and NVIDIA GPU Operator

  • Internal tests are being performed on Ubuntu 22.04.

Network Time Protocol

Nodes are required to be synchronized by time using NTP (Network Time Protocol) for proper system functionality.

Kubernetes Distribution

NVIDIA Run:ai control plane requires Kubernetes. The following Kubernetes distributions are supported:

  • Vanilla Kubernetes

  • NVIDIA Base Command Manager (BCM)

  • Elastic Kubernetes Engine (EKS)

  • Google Kubernetes Engine (GKE)

  • Azure Kubernetes Service (AKS)

  • Oracle Kubernetes Engine (OKE)

  • Rancher Kubernetes Engine 2 (RKE2)

See the following Kubernetes version support matrix for the latest NVIDIA Run:ai releases:

NVIDIA Run:ai version
Supported Kubernetes versions

v2.22

1.31 to 1.33

v2.23

1.31 to 1.34

v2.24 (latest)

1.33 to 1.35

For information on supported versions of managed Kubernetes, it's important to consult the release notes provided by your Kubernetes service provider. There, you can confirm the specific version of the underlying Kubernetes platform supported by the provider, ensuring compatibility with NVIDIA Run:ai. For an up-to-date end-of-life statement see Kubernetes Release Historyarrow-up-right or OpenShift Container Platform Life Cycle Policyarrow-up-right.

NVIDIA Run:ai Namespace

The NVIDIA Run:ai control plane uses a namespace runai-backend. Use the following to create the namespace:

Default Storage Class

The NVIDIA Run:ai control plane requires a default storage class to create persistent volume claims for NVIDIA Run:ai storage. The storage class, as per Kubernetes standards, controls the reclaim behavior, whether the NVIDIA Run:ai persistent data is saved or deleted when the NVIDIA Run:ai control plane is deleted.

circle-info

Note

For a simple (non-production) storage class example see Kubernetes Local Storage Classarrow-up-right. The storage class will set the directory /opt/local-path-provisioner to be used across all nodes as the path for provisioning persistent volumes. Then set the new storage class as default:

Kubernetes Ingress Controller

NVIDIA Run:ai cluster requires Kubernetes Ingress Controllerarrow-up-right to be installed on the Kubernetes cluster.

  • Make sure that a default ingress controller is set.

  • RKE2 comes with a pre-installed ingress controller.

There are multiple ways to install and configure an ingress controller. The following example demonstrates how to install and configure the HAProxyarrow-up-right ingress controller using helmarrow-up-right.

Run the following command:

Domain and Certificate Requirements

To install the NVIDIA Run:ai control plane in a NVIDIA Run:ai multi-tenant deployment, configure a wildcard DNS and wildcard certificate that allow secure access across tenant environments.

Wildcard DNS Record

In order to expose the NVIDIA Run:ai platform under a unified domain, configure a wildcard DNS record (e.g., *.runai.hostorg.com) that resolves to the cluster's load balancer IP address.

Wildcard TLS Certificate

You must provide a TLS certificate that matches your wildcard DNS domain (e.g., *.runai.hostorg.com). This certificate is used to secure HTTPS access to tenant-facing endpoints, ensuring that each tenant receives a secure URL when accessing NVIDIA Run:ai services.

Create a Kubernetes Secretarrow-up-right named runai-backend-tls in the runai-backend namespace and include the path to the TLS --cert and its corresponding private --key by running the following:

External Postgres Database (Optional)

The NVIDIA Run:ai control plane installation includes a default PostgreSQL database. However, you may opt to use an existing PostgreSQL database if you have specific requirements or preferences as detailed in External Postgres database configuration. Please ensure that your PostgreSQL database is version 16 or higher.

Last updated