2 Answers
- Newest
- Most votes
- Most comments
0
You can use AWS Batch for exporting and transforming teh 500TB of data from DynamoDB to an S3 bucket.
- Start by using the native export functionality of DynamoDB to export your data directly to an S3 bucket. This approach is highly efficient for large datasets and does not impact the performance of your DynamoDB table.
- Develop a Docker container with your transformation logic and upload it to Amazon ECR. Then, configure an AWS Batch environment specifying the necessary compute resources. Then, define job definitions in AWS Batch, detailing how jobs should run using your container. Last, submit transformation jobs to AWS Batch to process data from S3 and store the transformed data back to S3 or another location.
- Optionally, use AWS Step Functions to manage the workflow, particularly if the process involves multiple steps.
If this has resolved your issue or was helpful, accepting the answer would be greatly appreciated. Thank you!
0
Hi,
Yes, Glue is the ETL service on AWS for such tasks: it allows to process / transform data as you export from DDB to S3.
Here is a good article detailling how to do it: https://dev.to/ritaly/how-to-export-aws-dynamodb-data-to-s3-for-recurring-tasks-4l47
Best,
Didier
Relevant content
- asked 2 years ago
- asked 8 months ago
- asked 8 months ago
- asked 2 years ago
- AWS OFFICIALUpdated a month ago
- AWS OFFICIALUpdated a month ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated 3 months ago