- Newest
- Most votes
- Most comments
You have different options. The easiest I think would be to send the S3 events to an SQS queue with a VisibilityTimeout of 1 hour and a retention of 14 days. You will trigger the function from the queue. The function will try to upload the file to the API. If it failed, it will return a value indicating that the message processing failed. That message will become visible again in an hour and will be retried. This will happen up to 14 days, the maximum retention time for the queue.
If 14 days is not enough, you could also set up a DLQ with many retries. If it failed all those retries the message will be sent to the DLQ. You will then be able to re-drive them back to the original queue for another round.
There are other options for instance using StepFunctions. You will invoke a new state machine for each file. The state machine will call a Lambda function to upload the file. If upload is OK, state machine ends. If not, it gets into a loop with a Wait state. It can run for up to a year.
You could do it in a two step process: use the S3 event/lambda to create a record in a database (e.g. dynamodb) and then have another lambda that checks the database for files that need to be uploaded and does the upload. You can schedule that lambda to run every hour from a cloudwatch/eventbridge event.
Another thing to consider is using a deadletter queue on your current lambda and then replay the events on the deadletter queue via a scheduled lambda.
Relevant content
- AWS OFFICIALUpdated a month ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated 5 months ago