The error message in your log shows
Compile command returned: -9. This message typically indicates that the compiler process was killed. Normally this is due to the the OOM (out of memory) killer (run by the linux operating system) killing the compilation process due to memory exhaustion. The most recent version of torch-neuron should provide an updated message for
-9 errors that reflects the typical cause for this failure mode.
We recommend you try compiling on an instance with more memory, such as an inf1.6xlarge. Note: you only need the larger instance for compilation; you can still use a smaller instance (such as an inf1.xlarge) to run inference.
Please let us know if compiling on a larger instance resolved the error you’re seeing.
Export Autopilot model to GovCloud regionasked a month ago
Why does my kernal keep dying when I try to import Hugging Face BERT models to Amazon SageMaker?Accepted AnswerEXPERTasked 2 years ago
Why does my kernel keep restarting when I try to download pre-trained Hugging Face BERT models weights to Amazon SageMaker?asked 2 months ago
Deploy an model trained using Sagemaker's built-in k-NN algorithm for AWS Panoramaasked 6 months ago
Host a fine-tuned BERT Multilingual model on SageMaker with Serverless inferenceasked 6 months ago
How to register a multi container model to a model registry?asked 2 months ago
How to deploy pre-trained model?asked 3 years ago
Error: Sagemaker Lab for Hugging Face Exampleasked 3 months ago
Sagemaker training for multiclass classification run does not store the trained modelAccepted Answerasked 3 months ago
Not able to convert Hugging Face fine-tuned BERT model into AWS Neuronasked a month ago