AWS Bedrock batch inference stuck on SCHEDULED

0

Hello,

I’m trying to use Bedrock batch inference to process call transcripts at scale and have encountered an issue. In a small scale test with 100 records, each requiring a brief summary, the job status reaches "SCHEDULED" and remains there without progressing.

Since this is the only job currently running, I don’t believe it’s a quota issue. I'm using anthropic.claude-3-sonnet-20240229-v1:0 in us-east-1 with the following message format:

python

message = { "role": "user", "content": f"Summarize the following call transcript: {content}", } Additionally, one job I previously submitted completed with an error(the status was COMPLETED with errors in the output jsonl), suggesting that permissions may not be the issue either.

Could you provide guidance on what might be causing this delay or recommend steps to troubleshoot and resolve it?

Thank you!

已提問 3 個月前檢視次數 583 次
1 個回答
0

I understand you're experiencing an issue with your AWS Bedrock batch inference job getting stuck in the "SCHEDULED" status while processing call transcripts. This can be frustrating, but there are a few things we can consider to troubleshoot the problem.

First, even though you're not running multiple jobs concurrently, it's still worth verifying that you haven't hit any service quotas. Sometimes, quotas can be lower than expected, especially if you're using a new account or haven't requested increases.

Since a previous job completed (albeit with errors), it suggests that your basic permissions and setup are likely correct. However, the fact that it completed with errors indicates there might be an issue with the input data or the way the model is processing it.

Here are some steps you can take to troubleshoot:

  1. Check the job details in the AWS Management Console. Navigate to the Batch inference section and select your job to view its details. This might provide more information about why the job is stuck.

  2. Verify that your input data is correctly formatted. Ensure that each record in your JSONL file is properly structured with a unique identifier (recordId) and the modelInput in the format required by the Claude model.

  3. Consider reducing the batch size further for testing. Try with just a few records to see if the job progresses.

  4. Review the S3 bucket permissions to ensure that Bedrock has the necessary access to read the input and write the output.

  5. Check if there are any error messages or logs available in CloudWatch that might provide more insight into the issue.

  6. If possible, try running the same prompts through the Bedrock console or API directly (not as a batch job) to see if they process successfully. This can help isolate whether the issue is with the batch processing or the prompts themselves.

  7. Ensure that your account has the necessary permissions to use the Claude 3 Sonnet model in batch inference mode.

If none of these steps resolve the issue, it may be worth contacting AWS support for further assistance. They can look into the specifics of your account and job configuration to identify any underlying issues.

Remember, batch inference is designed to handle large-scale processing, so once you resolve this issue, it should provide an efficient way to summarize your call transcripts at scale.
Sources
Enhance call center efficiency using batch inference for transcript summarization with Amazon Bedrock | AWS Machine Learning Blog
Process multiple prompts with batch inference - Amazon Bedrock
Monitor batch inference jobs - Amazon Bedrock

profile picture
已回答 3 個月前
profile picture
專家
已審閱 2 個月前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南