Taking the first step
Purpose | Determine which AWS generative AI services are the best fit for your organization. |
Last updated | February 14, 2025 |
Covered services |
Introduction
Generative AI is a set of artificial intelligence (AI) systems and models designed to generate content such as code, text, images, music, or other forms of data. These systems can produce new content based on patterns and knowledge learned from existing data. Increasingly, organizations and businesses are using generative AI to:
-
Automate creative workflows — Use generative AI services to automate the workflows of time-consuming creative processes such as writing, image or video creation, and graphic design.
-
Customize and personalize content — Generate targeted content, product recommendations, and customized offerings for an audience-specific context.
-
Augment data — Synthesize large training datasets for other ML models to unlock scenarios where human-labeled data is scarce.
-
Reduce cost — Potentially lower costs by using synthesized data, content, and digital assets.
-
Faster experimentation — Test and iterate on more content variations and creative concepts than would be possible manually.
This guide helps you select the AWS generative AI services and tools that are the best fit for your needs and your organization.
A twelve-minute video about building generative AI applications on AWS, part one of a four-part series.
View part two
Understand
Amazon offers a range of generative AI services, applications, tools, and supporting infrastructure. Which of these you use depends a lot on the following factors:
-
What you’re trying to do
-
How much choice you need in the foundation models that you use
-
The degree of customization you need in your generative AI applications
-
The expertise within your organization

Amazon Q Business and Amazon Q Developer— Applications to boost productivity
At the top of Amazon's generative AI stack, Amazon Q Business and Amazon Q Developer use large language models (LLMs) and foundation models. However, they don’t require that you explicitly choose a model. Each of these applications support different use cases, and all are powered by Amazon Bedrock.
Learn more about the primary generative AI–powered assistants currently available:
Amazon Q Business can answer questions, provide summaries, generate content, and securely complete tasks based on the data in your enterprise systems. It supports the general use case of using generative AI to start making the most of the information in your enterprise. With Amazon Q Business, you can make English-language queries about that information. It provides responses in a manner appropriate to your team’s needs. In addition, you can create lightweight, purpose-built Amazon Q Apps within your Amazon Q Business Pro subscription.
Amazon Bedrock — Build and scale your generative AI application with FMs
If you're developing custom AI applications, need access to multiple foundation models, and
want more control over the AI models and outputs, then Amazon Bedrock could be the service that
meets your needs.
Amazon Bedrock
is a fully managed service for building generative AI applications, with support for
popular
foundation models, including Amazon Nova, Amazon Titan,
Anthropic Claude
Use the Amazon Bedrock Marketplace to discover, test, and use over 100 popular, emerging, and specialized FMs. Supported FMs are updated on a regular basis.

In addition, Amazon Bedrock provides what you need to build generative AI applications with security, privacy, and responsible AI—regardless of the foundation model you choose. It also offers model-independent, single API access and the flexibility to use different foundation models and upgrade to the latest model versions, with minimal code changes.
Learn more about the key features of Amazon Bedrock:
Model customization can deliver differentiated and personalized user experiences. To customize models for specific tasks, you can privately fine-tune FMs using your own labeled datasets. Custom models include capabilities such as fine-tuning and continued pre-training using unlabeled datasets. The list of FMs for which Amazon Bedrock supports fine-tuning includes Amazon Nova Micro, Lite, and Pro, Anthropic Claude 3 Haiku, Cohere Command, Meta Llama 2, Amazon Titan Text Lite and Express, Amazon Titan Multimodal Embeddings, and Amazon Titan Image Generator. The list of supported FMs is updated on an ongoing basis.
In addition, you can use Amazon Bedrock Custom Model Import to bring your own custom models and use them within Amazon Bedrock.
Amazon SageMaker AI (formerly Amazon SageMaker) — Build custom models and control the full ML lifecycle, from data preparation to model deployment and monitoring
With Amazon SageMaker AI, you can build, train, and deploy machine learning models, including FMs, at scale. Consider this option when you have use cases that can benefit from extensive training, fine-tuning, and customization of foundation models. It also streamlines the sometimes-challenging task of evaluating which FM is the best fit for your use case.
Amazon SageMaker AI also provides infrastructure and purpose-built tools for use throughout the ML lifecycle, including integrated development environments (IDEs), distributed training infrastructure, governance tools, machine learning operations (MLOps) tools, inference options and recommendations, and model evaluation.

Use Amazon SageMaker Partner AI Apps to access generative AI and machine learning (ML) development applications built, published, and distributed by industry-leading application providers. Partner AI Apps are certified to run on SageMaker AI. With Partner AI Apps, you can improve how you build solutions based on foundation models (FM) and classic ML models without compromising the security of your sensitive data. The data stays completely within your trusted security configuration and is never shared with a third party.
SageMaker AI is part of the next generation of Amazon SageMaker, which includes:
-
Amazon SageMaker Unified Studio (in preview), a unified development experience that brings together AWS data, analytics, artificial intelligence (AI), and machine learning (ML) services.
-
AWS tools for complete development workflows, including model development, generative AI app development, data processing, and SQL analytics, in a single governed environment. Create or join projects to collaborate with your teams, securely share AI and analytics artifacts, and access your data stored in Amazon Simple Storage Service (Amazon S3), Amazon Redshift, and more data sources through the Amazon SageMaker Lakehouse.
Explore key features of SageMaker AI that may help you determine when to use it:
Amazon SageMaker JumpStart is an ML hub that provides access to publicly available foundation models. Those models include Mistral, Llama 3, CodeLlama, and Falcon 2. They can be customized with advanced fine-tuning and deployment techniques such as Parameter Efficient Fine-Tuning (PEFT) and Low-Rank Adaptation (LoRA).
This following screenshot shows some of the available models in Amazon SageMaker JumpStart within the AWS Management Console.

Infrastructure to build and train AI models
AWS offers specialized, accelerated hardware for high performance ML training and inference.
-
AWS Trainium2-powered Amazon EC2 Trn2 instances
and Trn2 UltraServers deliver the highest performance for AI training and inference on AWS. -
AWS Trainium
is the second-generation ML accelerator that AWS has purpose-built for deep learning (DL) training of 100B+ parameter models. -
AWS Inferentia2-based Amazon EC2 Inf2 instances
are designed to deliver high performance at the lowest cost in Amazon EC2 for your DL and generative AI inference applications. -
Amazon EC2 P5 and P5e
instances, powered by NVIDIA H100 and NVIDIA H200 Tensor Core GPUs, which are well-suited for both training and inference tasks in machine learning. -
Amazon EC2 G6e
instances feature up to 8 NVIDIA L40S Tensor Core GPUs, and third generation AMD EPYC processors, for a wide range of graphics-intensive and machine learning use cases.
Consider
After you've decided on a generative AI service, choose the foundation model (FM) that gives you the best results for your use case.
Amazon Bedrock has a model evaluation capability that can assist in evaluating, comparing, and selecting the best FMs for your use case.
Here are some critical factors to consider when choosing an appropriate FM for your use case:
Identify use cases/modality
What it is: Modality refers to the type of data the model processes: text, images (vision), or embeddings.
Why it matters: The choice of modality should align with the data that you're working with. For example, if your project involves processing natural language, a text-based model like Claude, Llama 3.1, or Titan Text G1 is suitable. If you want to create embeddings, then you might use a model like Titan Embeddings G1. Similarly, for image-related tasks, models such as Stable Diffusion XL, and Titan Image Generator v2, are more appropriate. Your use case might also involve considering your data source and the support for data source connectors, such as those provided in Amazon Q Business.
Choose
Generative AI category | What is it optimized for? | Generative AI services |
---|---|---|
Amazon Q |
Generating code and providing responses to questions across business data by connecting to enterprise data repositories to summarize the data logically, analyze trends, and engage in dialogue about the data. |
|
Amazon Bedrock |
Offering a choice of foundation models, customizing them with your own data, and building generative AI applications with the builder tools that Amazon Bedrock offers. |
|
Amazon SageMaker AI |
Building, training, and deploying machine learning models, including foundation models, at scale. |
|
Infrastructure for FM training and inference |
Offering services that maximize the price performance benefits in FM training and inference. |
Use
Now that we've covered the criteria you need to apply in choosing an AWS generative AI service, you can select which services are optimized for your needs and explore how you might get started using each of them.
What is Amazon Q Business?
Get an overview of Amazon Q Business, with explanations of what it is, how it works, and how to get started using it.
Create a sample Amazon Q Business application
Learn how to create your first Amazon Q Business application in either the AWS Management Console or using the command line interface (CLI).
Combine Amazon Q Business and AWS IAM Identity Center to build generative AI apps
Build private and secure enterprise generative AI apps with Amazon Q Business and AWS IAM Identity Center.
Explore
Architecture diagrams
These reference architecture diagrams show examples of AWS AI and ML services in use.
Whitepapers
Explore whitepapers to help you get started and learn best practices in choosing and using AI and ML services.
AWS solutions
Explore vetted solutions and architectural guidance for common use cases for AI and ML services.
Resources
Public foundation models
Supported foundation models are updated on a regular basis, and currently include:
Use Amazon Bedrock and Amazon SageMaker AI to experiment with a variety of foundation models, and privately
customize them with your data. To explore generative AI quickly, you also have the option of
using PartyRock, an Amazon Bedrock Playground
Associated blog posts