Deploy a robust pipeline that uses managed automation tools and machine learning (ML) services to simplify ML model development and production
The ML lifecycle is an iterative and repetitive process that involves changing models over time and learning from new data. As ML applications gain popularity, organizations are building new and better applications for a wide range of use cases including optimized email campaigns, forecasting tools, recommendation engines, self-driving vehicles, virtual personal assistants, and more. While operational and pipelining processes vary across projects and organizations, the processes contain commonalities across use cases.
The MLOps Workload Orchestrator solution helps you streamline and enforce architecture best practices by providing an extendable framework for managing ML pipelines for Amazon Web Services (AWS) ML services and third-party services. The solution’s template allows you to train models, upload trained models, configure the pipeline orchestration, initiate the start of the deployment process, move models through different stages of deployment, and monitor the successes and failures of the operations. The solution also provides a pipeline for building and registering Docker images for custom algorithms that can be used for model deployment on an Amazon SageMaker AI
You can use batch and real-time data inferences to configure the pipelines for your business context. This solution increases your team’s agility and efficiency by allowing them to repeat successful processes at scale.
The MLOps Workload Orchestrator solution currently offers 12 pipelines:
-
One pipeline to train ML models using Amazon SageMaker AI built-in algorithms and Amazon SageMaker AI training job
-
One pipeline to train ML models using Amazon SageMaker AI built-in algorithms and Amazon SageMaker AI hyperparameter tuning job
-
One pipeline to train ML models using Amazon SageMaker AI built-in algorithms and Amazon SageMaker AI autopilot job
-
Two BYOM real-time inference pipelines for ML models trained using both Amazon SageMaker AI built-in algorithms and custom algorithms
-
Two BYOM batch transform pipelines for ML models trained using both Amazon SageMaker AI built-in algorithms and custom algorithms
-
One custom algorithm image builder pipeline that you can use to build and register Docker images in https://aws.amazon.com/ecr/
[Amazon Elastic Container Registry] (Amazon ECR) for custom algorithms -
Four model monitor pipelines to continuously monitor the quality of deployed machine learning models by the real-time inference pipeline and alerts for deviations in data quality, model quality, model bias, and/or model explainability
To support multiple use cases and business needs, this solution provides two AWS CloudFormation
-
Template option 1 - Single account - Use the single-account template to deploy all of the solution’s pipelines in the same AWS account. This option is suitable for experimentation, development, and/or small-scale production workloads.
-
Template option 2 - Multi-account - Use the multi-account template to provision multiple environments (for example, development, staging, and production) across different AWS accounts, which improves governance and increases security and control of the ML pipeline’s deployment, provides safe experimentation and faster innovation, and keeps production data and workloads secure and available to ensure business continuity.
This implementation guide provides an overview of the MLOps Workload Orchestrator solution, its reference architecture and components, considerations for planning the deployment, configuration steps for deploying the solution to the AWS Cloud.
The intended audience for using this solution’s features and capabilities in their environment includes IT infrastructure architects, ML engineers, data scientists, developers, DevOps, data analysts, and marketing technology professionals who have practical experience architecting in the AWS Cloud.
Use this navigation table to quickly find answers to these questions:
If you want to . . . | Read . . . |
---|---|
Know the cost for running this solution. The estimated cost for running this solution in the US East (N. Virginia) Region is USD $374.57 per month for AWS resources. |
|
Understand the security considerations for this solution. |
|
Know how to plan for quotas for this solution. |
|
Know which AWS Regions support this solution. |
|
View or download the AWS CloudFormation templates included in this solution to automatically deploy the infrastructure resources (the "stack") for this solution. |
|
Access the source code and optionally use the AWS Cloud Development Kit (AWS CDK) to deploy the solution. |