- Newest
- Most votes
- Most comments
Based on the information provided, it appears that the Meta Llama 3.2 3B model does not currently support on-demand throughput in Amazon Bedrock, which is why you're encountering the validation error when trying to create an application inference profile.
When you run the command to list foundation models that support ON_DEMAND inference, the Llama 3.2 3B model doesn't appear in the results because it doesn't support this inference type. This is consistent with the error message you're receiving.
Some models in Amazon Bedrock only support provisioned throughput rather than on-demand invocation. The Llama 3.2 models appear to fall into this category. This is why you were able to successfully create inference profiles for other models like Amazon Nova Lite, Nova Canva, Nova Pro, Nova Micro, and OpenAI GPT OSS, as these models do support on-demand inference.
To confirm which models support on-demand inference, you can use the ListFoundationModels API with the byInferenceType="ON_DEMAND" filter. This will return only those models that can be used with on-demand throughput.
Unfortunately, if the Llama 3.2 3B model doesn't support on-demand inference, you won't be able to create an application inference profile for it. You would need to use a different model that supports on-demand inference, or explore other options for using the Llama model with provisioned throughput if that's available.
Sources
Community | Find Amazon Bedrock models for immediate on-demand invocation
Community | Finding Amazon Bedrock models for immediate on-demand invocation
meta.llama3-2-11b-instruct-v1:0 not supported for on-demand throughput? | AWS re:Post
Relevant content
- asked 7 months ago
- asked a year ago
- AWS OFFICIALUpdated 5 months ago
