runai inference standard exec
execute a command in an inference workload
runai inference standard exec [WORKLOAD_NAME] [flags]Examples
# Execute bash in the workload
runai inference exec <workload-name> --tty --stdin -- /bin/bash
# Execute ls command in the workload
runai inference exec <workload-name> -- ls
# Execute a command in a specific pod of the workload
runai inference exec <workload-name> --pod <pod-name> -- nvidia-smiOptions
-c, --container string The name of the container within the pod.
-h, --help help for exec
--pod string The pod ID. If not specified, the first pod will be used.
--pod-running-timeout duration Timeout for pod to reach running state (e.g. 5s, 2m, 3h).
-p, --project string Specify the project for the command to use. Defaults to the project set in the context, if any. Use 'runai project set <project>' to set the default.
-i, --stdin Pass stdin to the container
-t, --tty Stdin is a TTY
--wait-timeout duration Timeout while waiting for the workload to become ready for log streaming (e.g., 5s, 2m, 3h).Options inherited from parent commands
SEE ALSO
runai inference standard - Runs a single inference process on one node. Suitable for smaller models or simpler inference tasks.
Last updated