Request Inferences from a Deployed Service (AWS CLI) - Amazon SageMaker AI

Request Inferences from a Deployed Service (AWS CLI)

Inference requests can be made with the sagemaker-runtime invoke-endpoint once you have an Amazon SageMaker AI endpoint InService. You can make inference requests with the AWS Command Line Interface (AWS CLI). The following example shows how to send an image for inference:

aws sagemaker-runtime invoke-endpoint --endpoint-name 'insert name of your endpoint here' --body fileb://image.jpg --content-type=application/x-image output_file.txt

An output_file.txt with information about your inference requests is made if the inference was successful.

For TensorFlow submit an input with application/json as the content type.

aws sagemaker-runtime invoke-endpoint --endpoint-name 'insert name of your endpoint here' --body fileb://input.json --content-type=application/json output_file.txt