Transfer from S3 to Redshift and filter specific key

0

Hey! My first task with AWS, so the question is more about the best way to solve the problem. The problem: I have integration from some service to export data to S3. After each export I need to transfer the data to Redshift and to filter some specific keys.

I checked those ways to do it:

  1. Lambda function to filter out the key
  2. Batch
  3. Transfer with airflow and filter the key by the list in the manifest

I’m checking now the 3rd option.

What do you think?

Alexgub
已提问 2 年前530 查看次数
2 回答
1

Hi, have you considered using Redshift Spectrum to import data into Redshift using SQL?

Assuming the data is in the format that can be read by Redshift Spectrum it would greatly simplify your ingestion pipeline.

You can schedule the query that does the import to automate the process https://docs.aws.amazon.com/redshift/latest/mgmt/query-editor-schedule-query.html

AWS
Alex_T
已回答 2 年前
1

You may want to consider AWS Glue, which is a serverless data integration service.

Load data from Amazon S3 to Amazon Redshift using AWS Glue - https://docs.aws.amazon.com/prescriptive-guidance/latest/patterns/load-data-from-amazon-s3-to-amazon-redshift-using-aws-glue.html

AWS
专家
Hernito
已回答 2 年前

您未登录。 登录 发布回答。

一个好的回答可以清楚地解答问题和提供建设性反馈,并能促进提问者的职业发展。

回答问题的准则