Select your cookie preferences

We use essential cookies and similar tools that are necessary to provide our site and services. We use performance cookies to collect anonymous statistics, so we can understand how customers use our site and make improvements. Essential cookies cannot be deactivated, but you can choose “Customize” or “Decline” to decline performance cookies.

If you agree, AWS and approved third parties will also use cookies to provide useful site features, remember your preferences, and display relevant content, including relevant advertising. To accept or decline all non-essential cookies, choose “Accept” or “Decline.” To make more detailed choices, choose “Customize.”

Model deployment options in Amazon SageMaker AI

Focus mode
Model deployment options in Amazon SageMaker AI - Amazon SageMaker AI

After you train your machine learning model, you can deploy it using Amazon SageMaker AI to get predictions. Amazon SageMaker AI supports the following ways to deploy a model, depending on your use case:

SageMaker AI also provides features to manage resources and optimize inference performance when deploying machine learning models:

  • To manage models on edge devices so that you can optimize, secure, monitor, and maintain machine learning models on fleets of edge devices, see Model deployment at the edge with SageMaker Edge Manager. This applies to edge devices like smart cameras, robots, personal computers, and mobile devices.

  • To optimize Gluon, Keras, MXNet, PyTorch, TensorFlow, TensorFlow-Lite, and ONNX models for inference on Android, Linux, and Windows machines based on processors from Ambarella, ARM, Intel, Nvidia, NXP, Qualcomm, Texas Instruments, and Xilinx, see Model performance optimization with SageMaker Neo.

For more information about all deployment options, see Deploy models for inference.

PrivacySite termsCookie preferences
© 2025, Amazon Web Services, Inc. or its affiliates. All rights reserved.