1 Risposta
- Più recenti
- Maggior numero di voti
- Maggior numero di commenti
0
Hi Seun,
What type of endpoint are you deploying and what error are you getting at inference time? Not sure if you are using Batch Transform. When using Batch inference, you have a maximum payload size of GBs and maximum runtime of days. If you're using another type of deployment, such as a real-time endpoint, you might be exceeding a limit of payload size or response timeout. See here for more details.
See this link for a simple script to see the cause of error in the Inference Recommender job. Inference Recommender is the easiest way to assess the minimum instance type that will suit the workload.
Hope this helps.
con risposta 7 mesi fa
Contenuto pertinente
- AWS UFFICIALEAggiornata 2 anni fa
- Perché il mio endpoint Amazon SageMaker entra in stato di errore quando creo o aggiorno un endpoint?AWS UFFICIALEAggiornata un anno fa
- AWS UFFICIALEAggiornata un anno fa