Reliably archiving many s3 files

0

I would like to be able to add many (hundreds of thousands) of small s3 files to a single archive elsewhere on s3. It doesn't necessarily need to be fast, but it does need to be reliable. I can stream data through an archiver and back to s3 in a single lambda on a small scale, but since I need to get every single object, at full scale it's a lot to ask from a single lambda.

Could I, for instance, use step function to run archiving lambdas against a subset of the files and perform multipart uploads into a single combined archive? Are there any better ways to achieve this sort of thing?

gefragt vor 2 Jahren275 Aufrufe
1 Antwort
0

If you can leverage something like Firehose in your application you can stream files to a data lake in S3 and then archive to glacier with lifecycle policies.

profile pictureAWS
EXPERTE
Rob_H
beantwortet vor 2 Jahren

Du bist nicht angemeldet. Anmelden um eine Antwort zu veröffentlichen.

Eine gute Antwort beantwortet die Frage klar, gibt konstruktives Feedback und fördert die berufliche Weiterentwicklung des Fragenstellers.

Richtlinien für die Beantwortung von Fragen