Stay up to date with the latest from the Knowledge Center. See all new Knowledge Center articles published in the last month, and re:Post's top contributors.
How do I optimize batch inference jobs in Amazon Bedrock?
1 minute read
3
My Amazon Bedrock batch inference jobs are slow or fail.
Resolution
Batch job execution times depend on available capacity, concurrent jobs in the queue, and model-specific resource allocation. Use the following resolution methods to optimize your batch inference jobs in Amazon Bedrock.
Provide simple and complete input prompts
To reduce the time to process the job and improve the quality of the results, create clear, concise prompts that don't include unnecessary context.
Don't exceed service quotas
If you run multiple batch inference jobs in parallel, then make sure that they don't exceed service quotas that vary by model and AWS Region.