How to Concurrent upload of docs to S3 and use SQS to process it

0

Hi,

We are basically trying to make use of SQS to process the upload request to S3 as we get simultaneously many requests for upload during the peak hours. As application is hosted on elastic beanstalk with auto scaling , it manages to an extent but we need to make the upload much faster and smooth typically file size ranges from 50MB to 100MB

Any ideas on infrastructure or ref’s to follow?

We came across the worker environment of elastic beanstalk but not sure if it’ll solve the problem. So we are also thinking to make a serverless architecture using lambda , SQS and S3

Any help really appreciated. Thanks

1개 답변
0

Hi,

S3 can easily manage concurrent uploads

You have to POST docs directly to "S3 Signed URLs" generated by your application

profile picture
답변함 일 년 전

로그인하지 않았습니다. 로그인해야 답변을 게시할 수 있습니다.

좋은 답변은 질문에 명확하게 답하고 건설적인 피드백을 제공하며 질문자의 전문적인 성장을 장려합니다.

질문 답변하기에 대한 가이드라인

관련 콘텐츠