awsbsub
Submits jobs to the job queue of the cluster.
awsbsub [-h] [-jn
JOB_NAME
] [-cCLUSTER
] [-cf] [-wWORKING_DIR
] [-pwPARENT_WORKING_DIR
] [-ifINPUT_FILE
] [-pVCPUS
] [-mMEMORY
] [-eENV
] [-ebENV_DENYLIST
] [-rRETRY_ATTEMPTS
] [-tTIMEOUT
] [-nNODES
] [-aARRAY_SIZE
] [-dDEPENDS_ON
] [command
] [arguments
[arguments
...]]
Important
AWS ParallelCluster doesn't support GPU jobs for AWS Batch. For more information, see GPU jobs.
Positional Arguments
command
-
Submits the job (the command specified must be available on the compute instances) or the file name to be transferred. See also
--command-file
. arguments
-
(Optional) Specifies arguments for the command or the command-file.
Named Arguments
-jn
JOB_NAME
, --job-nameJOB_NAME
-
Names the job. The first character must be either a letter or number. The job name can contain letters (both uppercase and lowercase), numbers, hyphens, and underscores, and be up to 128 characters in length.
-c
CLUSTER
, --clusterCLUSTER
-
Specifies the cluster to use.
-cf, --command-file
-
Indicates that the command is a file to be transferred to the compute instances.
Default: False
-w
WORKING_DIR
, --working-dirWORKING_DIR
-
Specifies the folder to use as the job's working directory. If a working directory isn't specified, the job is run in the
job-
subfolder of the user’s home directory. You can use either this parameter or the<AWS_BATCH_JOB_ID>
--parent-working-dir
parameter. -pw
PARENT_WORKING_DIR
, --parent-working-dirPARENT_WORKING_DIR
-
Specifies the parent folder of the job's working directory. If a parent working directory isn't specified, it defaults to the user’s home directory. A subfolder named
job-
is created in the parent working directory. You can use either this parameter or the<AWS_BATCH_JOB_ID>
--working-dir
parameter. -if
INPUT_FILE
, --input-fileINPUT_FILE
-
Specifies the file to be transferred to the compute instances, in the job's working directory. You can specify multiple input file parameters.
-p
VCPUS
, --vcpusVCPUS
-
Specifies the number of vCPUs to reserve for the container. When used together with
–nodes
, it identifies the number of vCPUs for each node.Default: 1
-m
MEMORY
, --memoryMEMORY
-
Specifies the hard limit of memory (in MiB) to provide for the job. If your job attempts to exceed the memory limit specified here, the job is ended.
Default: 128
-e
ENV
, --envENV
-
Specifies a comma-separated list of environment variable names to export to the job environment. To export all environment variables, specify ‘all’. Note that a list of 'all' environment variables doesn't include those listed in the
–env-blacklist
parameter, or variables starting with thePCLUSTER_*
orAWS_*
prefix. -eb
ENV_DENYLIST
, --env-blacklistENV_DENYLIST
-
Specifies a comma-separated list of environment variable names to not export to the job environment. By default,
HOME
,PWD
,USER
,PATH
,LD_LIBRARY_PATH
,TERM
, andTERMCAP
are not exported. -r
RETRY_ATTEMPTS
, --retry-attemptsRETRY_ATTEMPTS
-
Specifies the number of times to move a job to the
RUNNABLE
status. You can specify between 1 and 10 attempts. If the value of attempts is greater than 1, the job is retried if it fails, until it has moved to aRUNNABLE
status for the specified number of times.Default: 1
-t
TIMEOUT
, --timeoutTIMEOUT
-
Specifies the time duration in seconds (measured from the job attempt’s
startedAt
timestamp) after which AWS Batch terminates your job if it hasn't finished. The timeout value must be at least 60 seconds. -n
NODES
, --nodesNODES
-
Specifies the number of nodes to reserve for the job. Specify a value for this parameter to enable multi-node parallel submission.
Note
Multi-node parallel jobs aren't supported when the cluster_type parameter is set to
spot
. -a
ARRAY_SIZE
, --array-sizeARRAY_SIZE
-
Indicates the size of the array. You can specify a value between 2 and 10,000. If you specify array properties for a job, it becomes an array job.
-d
DEPENDS_ON
, --depends-onDEPENDS_ON
-
Specifies a semicolon-separated list of dependencies for a job. A job can depend upon a maximum of 20 jobs. You can specify a
SEQUENTIAL
type dependency without specifying a job ID for array jobs. A sequential dependency allows each child array job to complete sequentially, starting at index 0. You can also specify an N_TO_N type dependency with a job ID for array jobs. An N_TO_N dependency means that each index child of this job must wait for the corresponding index child of each dependency to complete before it can begin. The syntax for this parameter is "jobId=<string>
,type=<string>
;...".