# Installation

Run:ai is a Kubernetes-native orchestration and management platform designed to maximize GPU utilization for AI workloads.

## NVIDIA Run:ai System Components

NVIDIA Run:ai is made up of two components both installed over a [Kubernetes](https://kubernetes.io/) cluster:

* **NVIDIA Run:ai control plane** - Provides resource management, handles workload submission and provides cluster monitoring and analytics.
* **NVIDIA Run:ai cluster** - Provides enhanced scheduling and workload management, extending Kubernetes native capabilities.

As part of the installation process, you will:

* Create an NVIDIA Run:ai control plane tenant, provisioned and managed by NVIDIA
* Install one or more NVIDIA Run:ai clusters on your Kubernetes infrastructure and connect them to the tenant

<figure><img src="https://3278325112-files.gitbook.io/~/files/v0/b/gitbook-x-prod.appspot.com/o/spaces%2FLiY1aIqfxD3a58ufUYOM%2Fuploads%2FOZVB1cu98WakBHFPmsfi%2FSaaS.png?alt=media&#x26;token=e8c318d3-8131-4015-aaed-f20a9dd0075e" alt="" width="375"><figcaption></figcaption></figure>

## SaaS Deployment Model

The SaaS option is for organizations that prefer a fully managed control plane. With this model, NVIDIA hosts and operates the NVIDIA Run:ai control plane on your behalf. You are responsible only for installing the NVIDIA Run:ai cluster on your own Kubernetes infrastructure.

| Aspect        | Description                                                                     |
| ------------- | ------------------------------------------------------------------------------- |
| Control plane | Hosted and managed by NVIDIA. Access is provisioned via the NVIDIA NGC catalog. |
| Cluster       | Installed and managed by you on your Kubernetes infrastructure.                 |
| Connectivity  | The cluster connects to the NVIDIA-hosted control plane over the internet.      |
