Reliably archiving many s3 files

0

I would like to be able to add many (hundreds of thousands) of small s3 files to a single archive elsewhere on s3. It doesn't necessarily need to be fast, but it does need to be reliable. I can stream data through an archiver and back to s3 in a single lambda on a small scale, but since I need to get every single object, at full scale it's a lot to ask from a single lambda.

Could I, for instance, use step function to run archiving lambdas against a subset of the files and perform multipart uploads into a single combined archive? Are there any better ways to achieve this sort of thing?

已提問 2 年前檢視次數 278 次
1 個回答
0

If you can leverage something like Firehose in your application you can stream files to a data lake in S3 and then archive to glacier with lifecycle policies.

profile pictureAWS
專家
Rob_H
已回答 2 年前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南