Deploy an Inference Pipeline
An inference pipeline is an Amazon SageMaker model that is composed of a linear sequence of two to five containers that process requests for inferences on data. You use an inference pipeline to define and deploy any combination of pretrained Amazon SageMaker built-in algorithms and your own custom algorithms packaged in Docker containers. You can use an inference pipeline to combine preprocessing, predictions, and post-processing data science tasks. Inference pipelines are fully managed.
You can add Amazon SageMaker Spark ML Serving and scikit-learn containers that reuse the data transformers developed for training models. The entire assembled inference pipeline can be considered as an Amazon SageMaker model that you can use to make either real-time predictions or to process batch transforms directly without any external preprocessing.
Within an inference pipeline model, Amazon SageMaker handles invocations as a sequence of HTTP requests. The first container in the pipeline handles the initial request, then the intermediate response is sent as a request to the second container, and so on, for each container in the pipeline. Amazon SageMaker returns the final response to the client.
When you deploy the pipeline model, Amazon SageMaker installs and runs all of the
containers on each
Amazon Elastic Compute Cloud (Amazon EC2) instance in the endpoint or transform job.
Feature processing and
inferences run with low latency because the containers are co-located on the same
EC2
instances. You define the containers for a pipeline model using the CreateModel operation or from the
console. Instead of setting one PrimaryContainer, you use the
Containers parameter. to set the containers that make up the
pipeline You also specify the order in which the containers are executed.
A pipeline model is immutable, but you can update an inference pipeline by deploying a new one using the UpdateEndpoint operation. This modularity supports greater flexibility during experimentation.
There are no additional costs for using this feature. You pay only for the instances running on an endpoint.
Topics
Sample Notebooks for Inference Pipelines
For a sample notebook that uploads and processes a dataset, trains a model, and builds a pipeline model, see the Inference Pipelines with Spark ML and XGBoost on Abalone notebook. This notebook shows how you can build your machine learning pipeline by using Spark feature Transformers and the Amazon SageMaker XGBoost algorithm. After training the model, the sample shows how to deploy the pipeline (feature Transformer and XGBoost) for real-time predictions and also performs a batch transform job using the same pipeline.
For more examples that show how to create and deploy inference pipelines, see the Inference Pipelines with SparkML and BlazingText on DBPedia and Training using SparkML on EMR and hosting on SageMaker sample notebooks. For instructions on creating and accessing Jupyter notebook instances that you can use to run the example in Amazon SageMaker, see Use Notebook Instances.
To see a list of all the Amazon SageMaker samples, after creating and opening a notebook
instance,
choose the SageMaker Examples tab. There are three
inference pipeline notebooks. The first two inference pipeline notebooks just described
are located in the advanced_functionality folder and the third
notebook is in the sagemaker-python-sdk folder. To open a notebook,
choose its Use tab, then choose Create copy.
