Network requirements
Note
This section applies to self-hosted installation only. For SaaS network requirements, see the SaaS installation guide.
The following network requirements are for the NVIDIA Run:ai components installation and usage.
External access
Set out below are the domains to whitelist and ports to open for installation, upgrade, and usage of the application and its management.
Inbound rules
To allow your organization’s NVIDIA Run:ai users to interact with the cluster using the NVIDIA Run:ai Command-line interface, or access specific UI features, certain inbound ports need to be open.
NVIDIA Run:ai control plane
HTTPS entrypoint
0.0.0.0
NVIDIA Run:ai system nodes
443
NVIDIA Run:ai cluster
HTTPS entrypoint
0.0.0.0
NVIDIA Run:ai system nodes
443
Outbound rules
Note
Outbound rules applied to the NVIDIA Run:ai cluster component only, In case the NVIDIA Run:ai cluster is installed together with the NVIDIA Run:ai control plane - The NVIDIA Run:ai cluster FQDN refers to the NVIDIA Run:ai control plane FQDN.
For the NVIDIA Run:ai cluster installation and usage, certain outbound ports must be open.
Cluster sync
Sync NVIDIA Run:ai cluster with NVIDIA Run:ai control plane
NVIDIA Run:ai system nodes
NVIDIA Run:ai control plane FQDN
443
Metric store
Push NVIDIA Run:ai cluster metrics to NVIDIA Run:ai control plane's metric store
NVIDIA Run:ai system nodes
NVIDIA Run:ai control plane FQDN
443
Container Registry
Pull NVIDIA Run:ai images
All kubernetes nodes
runai.jfrog.io
443
Helm repository
NVIDIA Run:ai Helm repository for installation
Installer machine
runai.jfrog.io
443
The NVIDIA Run:ai installation has software requirements that require additional components to be installed on the cluster. This article includes simple installation examples which can be used optionally and require the following cluster outbound ports to be open:
Kubernetes Registry
Ingress Nginx image repository
All kubernetes nodes
registry.k8s.io
443
Google Container Registry
GPU Operator, and Knative image repository
All kubernetes nodes
gcr.io
443
Red Hat Container Registry
Prometheus Operator image repository
All kubernetes nodes
quay.io
443
Docker Hub Registry
Training Operator image repository
All kubernetes nodes
docker.io
443
Internal network
Ensure that all Kubernetes nodes can communicate with each other across all necessary ports. Kubernetes assumes full interconnectivity between nodes, so you must configure your network to allow this seamless communication. Specific port requirements may vary depending on your network setup.
Last updated