SageMaker AI Workflows - Amazon SageMaker AI

SageMaker AI Workflows

As you scale your machine learning (ML) operations, you can use Amazon SageMaker AI fully managed workflow services to implement continuous integration and deployment (CI/CD) practices for your ML lifecycle. With the Pipelines SDK, you choose and integrate pipeline steps into a unified solution that automates the model-building process from data preparation to model deployment. For Kubernetes based architectures, you can install SageMaker AI Operators on your Kubernetes cluster to create SageMaker AI jobs natively using the Kubernetes API and command-line Kubernetes tools such as kubectl. With SageMaker AI components for Kubeflow pipelines, you can create and monitor native SageMaker AI jobs from your Kubeflow Pipelines. The job parameters, status, and outputs from SageMaker AI are accessible from the Kubeflow Pipelines UI. Lastly, if you want to schedule non-interactive batch runs of your Jupyter notebook, use the notebook-based workflows service to initiate standalone or regular runs on a schedule you define.

In summary, SageMaker AI offers the following workflow technologies:

  • Pipelines: Tool for building and managing ML pipelines.

  • Kubernetes Orchestration: SageMaker AI custom operators for your Kubernetes cluster and components for Kubeflow Pipelines.

  • SageMaker Notebook Jobs: On demand or scheduled non-interactive batch runs of your Jupyter notebook.

You can also leverage other services that integrate with SageMaker AI to build your workflow. Options include the following services:

  • Airflow Workflows: SageMaker APIs to export configurations for creating and managing Airflow workflows.

  • AWS Step Functions: Multi-step ML workflows in Python that orchestrate SageMaker AI infrastructure without having to provision your resources separately.

For more information on managing SageMaker training and inference, see Amazon SageMaker Python SDK Workflows.