Reliably archiving many s3 files

0

I would like to be able to add many (hundreds of thousands) of small s3 files to a single archive elsewhere on s3. It doesn't necessarily need to be fast, but it does need to be reliable. I can stream data through an archiver and back to s3 in a single lambda on a small scale, but since I need to get every single object, at full scale it's a lot to ask from a single lambda.

Could I, for instance, use step function to run archiving lambdas against a subset of the files and perform multipart uploads into a single combined archive? Are there any better ways to achieve this sort of thing?

1 Risposta
0

If you can leverage something like Firehose in your application you can stream files to a data lake in S3 and then archive to glacier with lifecycle policies.

profile pictureAWS
ESPERTO
Rob_H
con risposta 2 anni fa

Accesso non effettuato. Accedi per postare una risposta.

Una buona risposta soddisfa chiaramente la domanda, fornisce un feedback costruttivo e incoraggia la crescita professionale del richiedente.

Linee guida per rispondere alle domande