Skip to content

Questions tagged with Meta Llama in Amazon Bedrock

Content language: English

Filter questions
Select tags to filter
Sort by
Sort by most recent
Filter Questions by:

Browse through the questions and answers listed below or filter and sort to narrow down your results.

6 results
I'm using invoke_model in Bedrock with Llama 4 Maverick. My prompt format looks like this (as per the docs): <|begin_of_text|> <|start_header_id|>system<|end_header_id|> ...system prompt...<|eot_id|...
2
answers
0
votes
37
views
asked 15 days ago
I’m trying to create an application inference profile for the Llama 3.2 3B model, but I’m getting the following error: **ValidationException: The provided foundation model does not support On Demand i...
1
answers
0
votes
57
views
asked a month ago
1
answers
0
votes
44
views
asked 2 months ago
I build 2 agents in Bedrock, Agent A is a Supervisor and Agent B analyzes data and returns a response. I started by using Claude Opus 4 for both agents but ran into a problem with cross region rate l...
1
answers
0
votes
90
views
asked 4 months ago
ARN loading from .env file is: arn:aws:bedrock:us-east-2:<Account-ID>:inference-profile/us.meta.llama4-scout-17b-instruct-v1:0 but getting error for diff ARN with diff zone & model-id `arn:aws:bedroc...
1
answers
0
votes
265
views
asked 6 months ago
I tried all available Meta models (some of them are not available), and when I read this page: https://aws.amazon.com/bedrock/meta/ , I can see some models have Max tokens of 128k, but when I use AWS ...
2
answers
0
votes
1.6K
views
asked 7 months ago
  • 1
  • Page size
    12 / page