Rekognition Bulk Analysis Info

0

Hello,

I would like to use Bulk Analysis info to all my entire bucket, meaning more than 10.000 documents but tasks are failing due to limit. So how can I achieve bulk processes into smaller batches? Can't I do 10.000, 10.000 for entire bucket?

I should be able to select entire bucket, if there is limit error it should give me error early of the process. If I can select entire bucket and can't more than 10.000 than I should be able to split.

How can I achieve my problem?

Thanks!

Alpcan
質問済み 5ヶ月前147ビュー
2回答
0

Hi Alpcan,

the 10000 limit is related to "Maximum number of images per Amazon Rekognition Media Analysis job", so you can have multiple parallel jobs for handling more than 10k images.

You can try to shard by folders or by file names ... Please look at this function as you might build upon it: https://github.com/aws-samples/amazon-rekognition-batch-processing-label-and-face-detection/blob/master/s3_detect_label_and_then_face.py

Note that you can have max 20 "Concurrent Amazon Rekognition Media Analysis jobs per account", that you can request to increase in the Service Quotas of your AWS account: https://us-east-1.console.aws.amazon.com/servicequotas/home/services/rekognition/quotas

Please accept the answer if it helped

profile pictureAWS
回答済み 5ヶ月前
profile picture
エキスパート
レビュー済み 5ヶ月前
0

I think I should be able to select entire bucket if I have option like 'bulk', it should automatically dive into 10k parts

Alpcan
回答済み 5ヶ月前

ログインしていません。 ログイン 回答を投稿する。

優れた回答とは、質問に明確に答え、建設的なフィードバックを提供し、質問者の専門分野におけるスキルの向上を促すものです。

質問に答えるためのガイドライン

関連するコンテンツ