1 個回答
- 最新
- 最多得票
- 最多評論
0
Hi Seun,
What type of endpoint are you deploying and what error are you getting at inference time? Not sure if you are using Batch Transform. When using Batch inference, you have a maximum payload size of GBs and maximum runtime of days. If you're using another type of deployment, such as a real-time endpoint, you might be exceeding a limit of payload size or response timeout. See here for more details.
See this link for a simple script to see the cause of error in the Inference Recommender job. Inference Recommender is the easiest way to assess the minimum instance type that will suit the workload.
Hope this helps.
已回答 7 個月前
相關內容
- 已提問 6 個月前
- 已提問 1 年前
- AWS 官方已更新 2 年前