Reliably archiving many s3 files

0

I would like to be able to add many (hundreds of thousands) of small s3 files to a single archive elsewhere on s3. It doesn't necessarily need to be fast, but it does need to be reliable. I can stream data through an archiver and back to s3 in a single lambda on a small scale, but since I need to get every single object, at full scale it's a lot to ask from a single lambda.

Could I, for instance, use step function to run archiving lambdas against a subset of the files and perform multipart uploads into a single combined archive? Are there any better ways to achieve this sort of thing?

已提问 2 年前264 查看次数
1 回答
0

If you can leverage something like Firehose in your application you can stream files to a data lake in S3 and then archive to glacier with lifecycle policies.

profile pictureAWS
专家
Rob_H
已回答 2 年前

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则