From the course: AWS Certified Machine Learning Engineer Associate (MLA-C01) Cert Prep
Unlock this course with a free trial
Join today to access over 24,700 courses taught by industry experts.
Multi-model vs. multi-container endpoints
From the course: AWS Certified Machine Learning Engineer Associate (MLA-C01) Cert Prep
Multi-model vs. multi-container endpoints
(gentle music) - [Instructor] Hello, guys. In today's lesson, we're going to talk about the multi-model and the multi-container deployments. So what is a multi-model endpoint? A multi-model endpoint allows a single SageMaker endpoint and a container image to dynamically load and serve multiple models, which optimizes both the cost and the efficiency. So you're sharing a single set of compute resources across multiple models, which significantly reduces the cost, and you're minimizing the complexity of managing individual endpoints for each model. So for the use cases, it could be suitable for when you have a large number of model that use the same machine learning framework, and they share similar code or logical flows. For example, you could deploy one model per user for personalized recommendations while hosting all the models on the same endpoint, and it could serve multiple variations of a language model tailored to different regions like English, French, or Spanish. So how does…
Practice while you learn with exercise files
Download the files the instructor uses to teach the course. Follow along and learn by watching, listening and practicing.
Contents
-
-
-
-
-
-
-
-
-
(Locked)
Intro: Model deployment53s
-
(Locked)
Online inference (real-time)20m 57s
-
(Locked)
Batch transform2m 17s
-
(Locked)
Other deployments8m 8s
-
(Locked)
Multi-model vs. multi-container endpoints10m 24s
-
(Locked)
Hands-on learning: Multi-model endpoint7m 16s
-
Hands-on learning: Multi-container endpoint2m 49s
-
(Locked)
SageMaker deployment7m 48s
-
(Locked)
Hands-on learning: XGBoost (churn prediction)6m 43s
-
(Locked)
Hands-on learning: Script mode3m 1s
-
(Locked)
Hands-on learning: Bring your own (BYO) Docker4m
-
(Locked)
SageMaker instance types3m 2s
-
(Locked)
SageMaker SDK7m 11s
-
(Locked)
Distributed training5m 20s
-
(Locked)
SageMaker Debugger3m 33s
-
Hands-on learning: SageMaker serverless inference6m 9s
-
(Locked)
SageMaker Autopilot3m 33s
-
(Locked)
Amazon SageMaker Inference Recommender6m 37s
-
(Locked)
Amazon SageMaker Serverless Inference5m 24s
-
(Locked)
Inference pipeline5m 3s
-
(Locked)
Hands-on learning: SageMaker Model Monitor15m 51s
-
(Locked)
SageMaker Neo6m 29s
-
(Locked)
SageMaker security6m 54s
-
(Locked)
Deployment target services10m 10s
-
(Locked)
Maintainable, scalable, cost-effective deployments8m 38s
-
(Locked)
Automatic scaling metrics4m 16s
-
(Locked)
Performance tradeoff analysis4m 10s
-
(Locked)
Apache Airflow, SageMaker Pipelines6m
-
(Locked)
Isolated ML system13m 12s
-
(Locked)
Exam cram11m 16s
-
(Locked)
-
-
-
-