The amazon-textract-transformer-pipeline sample shows a scalable batch PDF-to-images converter you might customize for this use case: Check out the code in the notebooks/preproc subfolder and the usage in the "Extract clean input images" section of notebook 1.
The current implementation is based on
poppler, pdf2image and processes batches of documents (from Amazon S3) through SageMaker Processing. It's probably not the most efficient possible (Python...), but can scale up to bigger instances (via multiprocessing) and out to multiple instances (via data sharding).
If you needed (near)-real-time processing instead of batch, you could probably get a similar solution running on a containerized Lambda function (poppler requires a lower-level install than pip). In our draft upgrade branch, we instead use SageMaker Asynchronous Inference for this... Because request/response payload sizes and memory could be theoretically very large for documents with many pages.
Images in Markdown fileasked 21 days ago
PDF to single JPEGs then Bytearrayasked 17 days ago
Does the Jenkins plugin for CodeBuild work with OpenShift provided Jenkins Docker images?Accepted Answerasked 3 years ago
API remote imagesasked 3 months ago
AWS Kendra - Search PDF with handwritten textAccepted Answerasked 3 months ago
Solution to delete new ECR images, from PutImage actions, that contain CRITICAL vulnerabilitesasked 13 days ago
Multiple images in a HIT with python APIasked 3 years ago
Are resized images served with cloudfront and resized with Lambda edge cached?Accepted Answerasked 3 days ago
FaceId's when indexing multiple face images of same personasked 4 months ago
Using Rekognition to process a batch of images on an S3 bucketasked 2 months ago