runai inference distributed scale
scale a distributed inference workload
runai inference distributed scale [WORKLOAD_NAME] [flags]Examples
# Scale a workload (replicas flag is required)
runai inference distributed scale <workload-name> --replicas 5Options
--dry-run If true, only print the object that would be sent, without sending it
-h, --help help for scale
-p, --project string Specify the project for the command to use. Defaults to the project set in the context, if any. Use 'runai project set <project>' to set the default.
--replicas int32 The number of replicas (sets of leader and workers) to run. Defaults to 1Options inherited from parent commands
--config-file string config file name; can be set by environment variable RUNAI_CLI_CONFIG_FILE (default "config.json")
--config-path string config path; can be set by environment variable RUNAI_CLI_CONFIG_PATH
-d, --debug enable debug mode
-q, --quiet enable quiet mode, suppress all output except error messages
--verbose enable verbose modeSEE ALSO
runai inference distributed - Runs multiple coordinated inference processes across multiple nodes. Required for models too large to run on a single node.
Last updated