Skip to content

All Content tagged with Meta Llama in Amazon Bedrock

Content language: English

Filter content
Select tags to filter
Sort by
Sort by most recent
6 results
I'm using invoke_model in Bedrock with Llama 4 Maverick. My prompt format looks like this (as per the docs): <|begin_of_text|> <|start_header_id|>system<|end_header_id|> ...system prompt...<|eot_id|...
2
answers
0
votes
44
views
asked 20 days ago
I’m trying to create an application inference profile for the Llama 3.2 3B model, but I’m getting the following error: **ValidationException: The provided foundation model does not support On Demand i...
1
answers
0
votes
65
views
asked a month ago
1
answers
0
votes
53
views
asked 2 months ago
I build 2 agents in Bedrock, Agent A is a Supervisor and Agent B analyzes data and returns a response. I started by using Claude Opus 4 for both agents but ran into a problem with cross region rate l...
1
answers
0
votes
93
views
asked 4 months ago
ARN loading from .env file is: arn:aws:bedrock:us-east-2:<Account-ID>:inference-profile/us.meta.llama4-scout-17b-instruct-v1:0 but getting error for diff ARN with diff zone & model-id `arn:aws:bedroc...
1
answers
0
votes
272
views
asked 6 months ago
I tried all available Meta models (some of them are not available), and when I read this page: https://aws.amazon.com/bedrock/meta/ , I can see some models have Max tokens of 128k, but when I use AWS ...
2
answers
0
votes
1.6K
views
asked 7 months ago
  • 1
  • Page size
    12 / page