Questions tagged with AWS Inferentia
Content language: English
Sort by most recent
Do ml.inf machines support multi-model endpoints?
We have been trying to deploy our multiple models to a multi-model endpoint that uses inference machines (inf.xlarge) without luck. ClientError: An error occurred (ValidationException) when calling the CreateEndpointConfig operation: MultiModel mode is not supported for instance type ml.inf1.xlarge. This isn't good, is that really the case, or have we messed up somewhere during the process? Thanks