ModelInvocationJobSummary
A summary of a batch inference job.
Contents
- inputDataConfig
-
Details about the location of the input to the batch inference job.
Type: ModelInvocationJobInputDataConfig object
Note: This object is a Union. Only one member of this object can be specified or returned.
Required: Yes
- jobArn
-
The Amazon Resource Name (ARN) of the batch inference job.
Type: String
Length Constraints: Minimum length of 0. Maximum length of 1011.
Pattern:
^(arn:aws(-[^:]+)?:bedrock:[a-z0-9-]{1,20}:[0-9]{12}:model-invocation-job/[a-z0-9]{12})$
Required: Yes
- jobName
-
The name of the batch inference job.
Type: String
Length Constraints: Minimum length of 1. Maximum length of 63.
Pattern:
^[a-zA-Z0-9]{1,63}(-*[a-zA-Z0-9\+\-\.]){0,63}$
Required: Yes
- modelId
-
The unique identifier of the foundation model used for model inference.
Type: String
Length Constraints: Minimum length of 1. Maximum length of 2048.
Pattern:
^(arn:aws(-[^:]+)?:bedrock:[a-z0-9-]{1,20}:(([0-9]{12}:custom-model/[a-z0-9-]{1,63}[.]{1}[a-z0-9-:]{1,63}/[a-z0-9]{12}$)|(:foundation-model/[a-z0-9-]{1,63}[.]{1}[a-z0-9-]{1,63}$)))|([a-z0-9-]{1,63}[.]{1}[a-z0-9-]{1,63}([.]?[a-z0-9-]{1,63})([:][a-z0-9-]{1,63}){0,2})|(([0-9a-zA-Z][_-]?)+)$
Required: Yes
- outputDataConfig
-
Details about the location of the output of the batch inference job.
Type: ModelInvocationJobOutputDataConfig object
Note: This object is a Union. Only one member of this object can be specified or returned.
Required: Yes
- roleArn
-
The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.
Type: String
Length Constraints: Minimum length of 0. Maximum length of 2048.
Pattern:
^arn:aws(-[^:]+)?:iam::([0-9]{12})?:role/.+$
Required: Yes
- submitTime
-
The time at which the batch inference job was submitted.
Type: Timestamp
Required: Yes
- clientRequestToken
-
A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.
Type: String
Length Constraints: Minimum length of 1. Maximum length of 256.
Pattern:
^[a-zA-Z0-9]{1,256}(-*[a-zA-Z0-9]){0,256}$
Required: No
- endTime
-
The time at which the batch inference job ended.
Type: Timestamp
Required: No
- jobExpirationTime
-
The time at which the batch inference job times or timed out.
Type: Timestamp
Required: No
- lastModifiedTime
-
The time at which the batch inference job was last modified.
Type: Timestamp
Required: No
- message
-
If the batch inference job failed, this field contains a message describing why the job failed.
Type: String
Length Constraints: Minimum length of 0. Maximum length of 2048.
Required: No
- status
-
The status of the batch inference job.
The following statuses are possible:
-
Submitted – This job has been submitted to a queue for validation.
-
Validating – This job is being validated for the requirements described in Format and upload your batch inference data. The criteria include the following:
-
Your IAM service role has access to the Amazon S3 buckets containing your files.
-
Your files are .jsonl files and each individual record is a JSON object in the correct format. Note that validation doesn't check if the
modelInput
value matches the request body for the model. -
Your files fulfill the requirements for file size and number of records. For more information, see Quotas for Amazon Bedrock.
-
-
Scheduled – This job has been validated and is now in a queue. The job will automatically start when it reaches its turn.
-
Expired – This job timed out because it was scheduled but didn't begin before the set timeout duration. Submit a new job request.
-
InProgress – This job has begun. You can start viewing the results in the output S3 location.
-
Completed – This job has successfully completed. View the output files in the output S3 location.
-
PartiallyCompleted – This job has partially completed. Not all of your records could be processed in time. View the output files in the output S3 location.
-
Failed – This job has failed. Check the failure message for any further details. For further assistance, reach out to the AWS Support Center
. -
Stopped – This job was stopped by a user.
-
Stopping – This job is being stopped by a user.
Type: String
Valid Values:
Submitted | InProgress | Completed | Failed | Stopping | Stopped | PartiallyCompleted | Expired | Validating | Scheduled
Required: No
-
- timeoutDurationInHours
-
The number of hours after which the batch inference job was set to time out.
Type: Integer
Valid Range: Minimum value of 24. Maximum value of 168.
Required: No
- vpcConfig
-
The configuration of the Virtual Private Cloud (VPC) for the data in the batch inference job. For more information, see Protect batch inference jobs using a VPC.
Type: VpcConfig object
Required: No
See Also
For more information about using this API in one of the language-specific AWS SDKs, see the following: