1 réponse
- Le plus récent
- Le plus de votes
- La plupart des commentaires
2
To my knowledge, no - it's not generally possible to pull the built-in algorithm containers outside SageMaker: Your easiest route would probably just be to deploy the model on SageMaker and integrate your other containerized tasks to call the SageMaker endpoint.
It's maybe worth mentioning that the framework containers for custom/script-mode modelling (e.g. the AWS DLCs for PyTorch/HuggingFace/etc) are not subject to this restriction (can check you should even be able to pull them locally): So if you were to use those to implement a customized text processing model I think you should be able to deploy it on ECS if needed. Of course this'd mean a more initial build and later maintenance effort though.
Contenus pertinents
- demandé il y a un an
- demandé il y a 4 mois
- demandé il y a 7 mois
- demandé il y a un an
- AWS OFFICIELA mis à jour il y a 2 ans
- AWS OFFICIELA mis à jour il y a 5 ans
- AWS OFFICIELA mis à jour il y a 8 mois
- AWS OFFICIELA mis à jour il y a 5 ans
Thank you so much Alex. I also found deep learning containers(DLC), but there was no container image for the BlazingText algorithm with inference code. The reason why I want to deploy BlazingText model on Fargate is to elastically respond to the varying workload on model endpoint. I think there are two options for my case: 1) selecting a proper instance type with auto scailing, 2) sagemaker serverless endpoint(preview).