multi model endpoints in sagemaker?

0

is there a list of containers documented somewhere , that support multi model endpoints ?

已提问 1 年前251 查看次数
2 回答
2

maybe this list can help you

https://docs.aws.amazon.com/sagemaker/latest/dg/sagemaker-algo-docker-registry-paths.html

first, choose your region then elect your algorithm then it will show you the docker registry

profile picture
专家
已回答 1 年前
0

Hi,

For information about the algorithms, frameworks, and instance types that you can use with multi-model endpoints, see the official document. Note that it has both cpu and gpu back-end instances. You can always bring your own if not already supported.

Hope it helps.

AWS
已回答 1 年前
  • @Tina_Qian - thanks, i went through the documentation, but one thing is not clear to me, are we compressing two trained models into one zipped/tar file?

  • Actually not. Basically, the SageMaker manages the lifecycle of models hosted on multi-model endpoints in the container's memory. Instead of downloading all of the models from an Amazon S3 bucket to the container when you create the endpoint, SageMaker dynamically loads and caches them when you invoke them. When SageMaker receives an invocation request for a particular model, it first routes the request to an instance behind the endpoint, then downloads the model from the S3 bucket to that instance's storage volume. Finally, it loads the model to the container's memory (CPU or GPU, depending on whether you have CPU or GPU backed instances) on that accelerated compute instance. If the model is already loaded in the container's memory, invocation is faster because SageMaker doesn't need to download and load it.

    Maybe the pictures from https://aws.amazon.com/cn/blogs/machine-learning/save-on-inference-costs-by-using-amazon-sagemaker-multi-model-endpoints/ is easy to follow and understand.

    Hope it helps.

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则