NVIDIA Run:ai REST API Documentation

NVIDIA Run:ai is a platform for accelerating and managing AI workloads in shared computing environments. It enables teams to dynamically schedule, allocate, and monitor GPU resources across clusters.

The NVIDIA Run:ai control plane API provides programmatic access to these capabilities, allowing you to integrate workload submission, resource management, and administrative operations directly into your own applications, services, or automation pipelines.

Why Integrate with NVIDIA Run:ai APIs

Developers use the NVIDIA Run:ai APIs to:

  • Automate workload management - Submit, suspend/resume, and monitor AI training and inference jobs programmatically.

  • Optimize resource utilization - Dynamically allocate GPU and CPU resources based on workload demand.

  • Integrate with existing systems - Connect NVIDIA Run:ai with CI/CD pipelines, data science platforms, or custom dashboards.

  • Manage multi-tenant environments - Streamline access control, quotas, and user management across teams or projects.

By integrating with the API, you gain fine-grained control and scalability beyond the user interface, making it easier to embed NVIDIA Run:ai capabilities into your organization’s AI infrastructure.

Last updated