Questions tagged with Extract Transform & Load Data
Content language: English
Select up to 5 tags to filter
Sort by most recent
Browse through the questions and answers listed below or filter and sort to narrow down your results.
hi team, can I ask why Glue is generating so many parquet files from my ETL job?
![Enter image description here](/media/postImages/original/IM6V7UVsE-QSi5AEKRNdOqkQ)
![Enter image description...
2
answers
0
votes
391
views
I am using AWS Glue and using the Glue Console to create ETL jobs for data transfer between Salesforce and AWS S3 bucket. I am using third party (Progress DataDirect and CData) connectors to connect...
1
answers
0
votes
295
views
asked 8 months agolg...
Our current setup involves AWS Glue in operation, where data is being extracted from one SQL Server and loaded into another SQL Server through use of AWS Glue Studio for selected tables.
Is there a...
1
answers
0
votes
192
views
asked 9 months agolg...
How do I connect Amazon RDS - Microsoft SQL Server through Glue Spark type jobs using python?
1
answers
0
votes
581
views
asked 9 months agolg...
I want to create a glue job to process multiple tables in parallel. If all the tables are to be processed in the same manner, is it possible to do it in only one glue job?
1
answers
0
votes
384
views
asked 9 months agolg...
I have a CSV file delivered by external vendor, to S3 and this file has some Non-ASCII/Junk characters. Before loading this to redahft table, I will need to remove these characters. I tried TRIMBLANKS...
1
answers
0
votes
397
views
asked 9 months agolg...
I was trying to perform Glue ETL transformation and store it in AWS Serverless Redshift database and S3 (both) . However, even the Console generated PySpark sheet fails. Almost none of the methods...
0
answers
0
votes
168
views
asked 9 months agolg...
I'm seeking guidance on a specific requirement and the recommended approach to achieve it.
My data is currently stored in an AWS Aurora SQL database (let's say host1/db1) . The objective is to...
2
answers
0
votes
370
views
asked 9 months agolg...
Hi
I have an architecture like below
user upload file -> S3 -> lambda trigger glue job -> glue job pull the file, read content, and save to a record in a table in Aurora Postgres
Everything is...
1
answers
0
votes
422
views
asked 9 months agolg...
We are in the process of importing data. The data will be provided as flat files or csv and stored in an Amazon S3 bucket. Each file is expected to be approximately 2GB in size with around 200k...
1
answers
0
votes
424
views
asked 9 months agolg...
Hi,
The environment is, there are multiple JSON files in a S3 bucket. I would like to add all of them to the Athena Table with the filtered values. I used the ChatGPT for the Athena query to create...
2
answers
0
votes
2941
views
asked 9 months agolg...
Hello All Experts,
Please help with the below scenario.
Data is stored in the raw zone and a column "ga4_dt "is extracted as a string in the format 'yyyymmdd' example
20230108. I can't update the...
2
answers
0
votes
1176
views
asked 9 months agolg...