Select your cookie preferences

We use essential cookies and similar tools that are necessary to provide our site and services. We use performance cookies to collect anonymous statistics, so we can understand how customers use our site and make improvements. Essential cookies cannot be deactivated, but you can choose “Customize” or “Decline” to decline performance cookies.

If you agree, AWS and approved third parties will also use cookies to provide useful site features, remember your preferences, and display relevant content, including relevant advertising. To accept or decline all non-essential cookies, choose “Accept” or “Decline.” To make more detailed choices, choose “Customize.”

Submit prompts and generate responses with model inference

Focus mode
Submit prompts and generate responses with model inference - Amazon Bedrock

Inference refers to the process of generating an output from an input provided to a model.

Amazon Bedrock offers a suite of foundation models that you can use to generate outputs of the following modalities. To see modality support by foundation model, refer to Supported foundation models in Amazon Bedrock.

Output modality Description Example use cases
Text Provide text input and generate various types of text Chat, question-and-answering, brainstorming, summarization, code generation, table creation, data formatting, rewriting
Image Provide text or input images and generate or modify images Image generation, image editing, image variation
Video Provide text or reference images and generate a video Video generation, image conversion to video
Embeddings Provide text, images, or both text and images and generate a vector of numeric values that represent the input. The output vector can be compared to other embeddings vectors to determine semantic similarity (for text) or visual similarity (for images). Text and image search, query, categorization, recommendations, personalization, knowledge base creation

You can directly run model inference in the following ways:

The following Amazon Bedrock features also use model inference as a step in a larger workflow:

  • Model evaluation uses the model invocation process to evaluate the performance of different models after you submit a CreateEvaluationJob request.

  • Knowledge bases use model invocation when using the RetrieveAndGenerate API to generate a response based on results retrieved from a knowledge base.

  • Agents use model invocation to generate responses in various stages during an InvokeAgent request.

  • Flows include Amazon Bedrock resources, such as prompts, knowledge bases, and agents, which use model invocation.

After testing out different foundation models with different prompts and inference parameters, you can configure your application to call these APIs with your desired specifications.

PrivacySite termsCookie preferences
© 2025, Amazon Web Services, Inc. or its affiliates. All rights reserved.