GPU search time for SageMaker Async Inference and general GPU availability.

0

Background I want to build an ML Inference pipeline that will use SageMaker Asynchronous Inference. To decrease costs I want to down all SageMaker Async Inference-related EC2s when no jobs are waiting (for example for time out of business hours or during working hours where there are no requests from my users).

The questions

  1. On average, how long does it take for AWS SageMaker Async Inference to get an up-and-running EC2 with a GPU ready to execute my ML tasks/inference?
  2. What is the current availability of GPU machines on AWS? Is there any shortage?
已提問 2 個月前檢視次數 74 次
沒有答案

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南