Combine streams from S3 object listing and ObjectCreated events

0

I am hoping to create a stream of messages, and the stream will be defined by providing an S3 bucket + prefix, and will stream:

  1. a list of all existing objects at the given prefix
  2. new objects that are uploaded under the prefix, via the ObjectCreated event

It will need to handle deduplication, so that if the same object appears in both 1 and 2 (due to async processing), the file will be processed exactly once.

I need to be able to create new instances of the stream at any time, and direct them to destinations such as Lambda or SQS, e.g. when changing the code that processes the file to produce a newer version of the output, I will create a new Lambda and replay the stream from the beginning (with a preference for newer files first)

Currently I am using the Databricks Autoloader, but I would like to move my solution back to AWS to have more control over the implementation. How can I build this functionality on AWS? Thanks!

For context, I have also discussed this solution in the Databricks forums: https://community.databricks.com/s/question/0D58Y000091yskKSAQ/help-integrating-streaming-pipeline-with-aws-services-s3-and-lambda

沒有答案

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南