Manage a batch inference job - Amazon Bedrock

Manage a batch inference job

After you create a batch inference job, you can view details about it, monitor its status, or stop it if it hasn't yet completed.

View details about a batch inference job

Apart from the configurations you set for a batch inference job, you can also monitor its progress by seeing how many records have been processed and how many records failed to process. To learn how to view details about batch inference jobs, select the tab corresponding to your method of choice and follow the steps.

Console
To view information about batch inference jobs
  1. Sign in to the AWS Management Console using an IAM role with Amazon Bedrock permissions, and open the Amazon Bedrock console at https://console.aws.amazon.com/bedrock/.

  2. From the left navigation pane, select Batch inference.

  3. In the Batch inference jobs section, choose a job.

  4. On the job details page, you can view information about the job's configuration and monitor its progress in the Processed records and Failed records field.

API

To get information about a batch inference job, send a GetModelInvocationJob request (see link for request and response formats and field details) with an Amazon Bedrock control plane endpoint and provide the ID or ARN of the job in the jobIdentifier field.

To list information about multiple batch inference jobs, send ListModelInvocationJobs request (see link for request and response formats and field details) with an Amazon Bedrock control plane endpoint. You can specify the following optional parameters:

Field Short description
maxResults The maximum number of results to return in a response.
nextToken If there are more results than the number you specified in the maxResults field, the response returns a nextToken value. To see the next batch of results, send the nextToken value in another request.

To list all the tags for a job, send a ListTagsForResource request (see link for request and response formats and field details) with an Amazon Bedrock control plane endpoint and include the Amazon Resource Name (ARN) of the job.

Stop a batch inference job

To learn how to stop an ongoing batch inference job, select the tab corresponding to your method of choice and follow the steps..

Console
To stop a batch inference job
  1. Sign in to the AWS Management Console using an IAM role with Amazon Bedrock permissions, and open the Amazon Bedrock console at https://console.aws.amazon.com/bedrock/.

  2. From the left navigation pane, select Batch inference.

  3. Select a job to go to the job details page or select the option button next to a job.

  4. Choose Stop job.

  5. Review the message and choose Stop job to confirm.

    Note

    You're charged for tokens that have already been processed.

API

To stop a batch inference job, send a StopModelInvocationJob request (see link for request and response formats and field details) with an Amazon Bedrock control plane endpoint and provide the ID or ARN of the job in the jobIdentifier field.

If the job was successfully stopped, you receive an HTTP 200 response.