Network Requirements
NVIDIA Run:ai requires certain network connectivity and access. This section outlines the network endpoints and protocols that must be reachable from your NVIDIA Run:ai control plane and cluster nodes to support installation, artifact retrieval, and ongoing platform communication.
Meeting these network requirements ensures that:
Clusters can register with and communicate to the control plane
The platform can access external services required for monitoring, logging, and artifact distribution
Follow the guidance below to verify and configure network access before proceeding with installation.
External Access
Listed below are the domains to whitelist and ports to open for installation, upgrades, and usage of the application and its management.
Note
Ensure the inbound and outbound rules are correctly applied to your firewall.
Inbound Rules
To allow your organization’s NVIDIA Run:ai users to interact with the cluster using the NVIDIA Run:ai Command-line interface, or access specific UI features, certain inbound ports need to be open.
NVIDIA Run:ai cluster
HTTPS entrypoint
0.0.0.0
NVIDIA Run:ai system nodes
443
Outbound Rules
For the NVIDIA Run:ai cluster installation and usage, certain outbound ports must be open:
Cluster sync
Sync NVIDIA Run:ai cluster with NVIDIA Run:ai control plane
NVIDIA Run:ai cluster system nodes
NVIDIA Run:ai control plane FQDN
443
Metric store
Push NVIDIA Run:ai cluster metrics to NVIDIA Run:ai control plane's metric store
NVIDIA Run:ai cluster system nodes
NVIDIA Run:ai control plane FQDN
443
Container Registry
Pull NVIDIA Run:ai images and Helm chart for installation
All kubernetes nodes
runai.jfrog.io and JFrog Cloud Storage URLs
443
Helm repository
NVIDIA Run:ai Helm repository for installation
Installer machine
runai.jfrog.io
443
The NVIDIA Run:ai installation has software requirements that require additional components to be installed on the cluster. This article includes simple installation examples which can be used optionally and require the following cluster outbound ports to be open:
Kubernetes Registry
Ingress HAProxy image repository
All kubernetes nodes
docker.io
443
Google Container Registry
GPU Operator, and Knative image repository
All kubernetes nodes
gcr.io
443
Red Hat Container Registry
Prometheus Operator image repository
All kubernetes nodes
quay.io
443
Docker Hub Registry
Training Operator image repository
All kubernetes nodes
docker.io
443
Internal Network
Ensure that all Kubernetes nodes can communicate with each other across all necessary ports. Kubernetes assumes full interconnectivity between nodes, so you must configure your network to allow this seamless communication. Specific port requirements may vary depending on your network setup.
Last updated