/AWS1/CL_SGM=>CREATEINFERENCECOMPONENT()
¶
About CreateInferenceComponent¶
Creates an inference component, which is a SageMaker hosting object that you can use to deploy a model to an endpoint. In the inference component settings, you specify the model, the endpoint, and how the model utilizes the resources that the endpoint hosts. You can optimize resource utilization by tailoring how the required CPU cores, accelerators, and memory are allocated. You can deploy multiple inference components to an endpoint, where each inference component contains one model and the resource utilization needs for that individual model. After you deploy an inference component, you can directly invoke the associated model when you use the InvokeEndpoint API action.
Method Signature¶
IMPORTING¶
Required arguments:¶
IV_INFERENCECOMPONENTNAME
TYPE /AWS1/SGMINFERENCECOMPONENTN00
/AWS1/SGMINFERENCECOMPONENTN00
¶
A unique name to assign to the inference component.
IV_ENDPOINTNAME
TYPE /AWS1/SGMENDPOINTNAME
/AWS1/SGMENDPOINTNAME
¶
The name of an existing endpoint where you host the inference component.
IV_VARIANTNAME
TYPE /AWS1/SGMVARIANTNAME
/AWS1/SGMVARIANTNAME
¶
The name of an existing production variant where you host the inference component.
IO_SPECIFICATION
TYPE REF TO /AWS1/CL_SGMINFERENCECOMPONE00
/AWS1/CL_SGMINFERENCECOMPONE00
¶
Details about the resources to deploy with this inference component, including the model, container, and compute resources.
IO_RUNTIMECONFIG
TYPE REF TO /AWS1/CL_SGMINFERENCECOMPONE04
/AWS1/CL_SGMINFERENCECOMPONE04
¶
Runtime settings for a model that is deployed with an inference component.
Optional arguments:¶
IT_TAGS
TYPE /AWS1/CL_SGMTAG=>TT_TAGLIST
TT_TAGLIST
¶
A list of key-value pairs associated with the model. For more information, see Tagging Amazon Web Services resources in the Amazon Web Services General Reference.