2 Answers
- Newest
- Most votes
- Most comments
0
Have you already checked this AWS documentation. As per this documentation:
Hope this helps.
Abhishek
I thought "max_tokens_to_sample" only affected the response. Anyway, I played around with that parameter and set it all the way to 30K, in which my query actually worked. It's odd since the documentation says the limit is 8K. I could not set it above 30K however.
0
Also having this issue. When we support more than about 10k tokens we get "An error occurred (ValidationException) when calling the InvokeModel operation: Input is too long for requested model.". Well below the 100k token limit we get through Anthropic's API. Also getting an error when posting the same request in the playground. I think something's actually broken in Bedrock
answered 6 months ago
Relevant content
- asked 2 months ago
- asked 4 months ago
- Accepted Answerasked 5 months ago
- AWS OFFICIALUpdated 3 months ago
- AWS OFFICIALUpdated 3 months ago
- AWS OFFICIALUpdated 10 months ago
- AWS OFFICIALUpdated a month ago
I too am getting a consistent error of - The read operation timed out. AWSHTTPSConnectionPool(host='bedrock-runtime.us-east-1.amazonaws.com', port=443): Read timed out. (read timeout=60). This occurs when the document text which I am sending exceeds 20K words. On trying with different values to max_tokens_to_sample parameter, api is throwing throttle exception. Any leads on how to proceed? Thanks in advance.
I'm also seeing performance way below the indicated 100k/200k token limits. Using word count as an approximate token proxy, I can get maybe 20k words maximum.
I get request timeouts with a body size of around 200k characters.