Questions tagged with Extract Transform & Load Data
Content language: English
Select up to 5 tags to filter
Sort by most recent
Browse through the questions and answers listed below or filter and sort to narrow down your results.
Why doesn't Glue Job and Glue Workflow have the function of version control and alias likes Labmda.lg...
I tried to develop the data orchestlation with s3, Glue Job and Glue Workflow. After I developed it, I found that Glue Job and Glue Workflow doesn't have the function of version control and alias...
0
answers
0
votes
28
views
asked 16 hours agolg...
Hi team, first post, let me know if it provides a good explanation.
I'd like to know a way to minimize the effort for data ingestion.
We have two options as follows:
(1) csv files from a file...
0
answers
0
votes
32
views
asked a day agolg...
I am running an EMR cluster with an attached notebook, and using Apache spark to load/process data however I have not been able to load data into Apache. Whenever I try to run...
1
answers
0
votes
35
views
asked 3 days agolg...
Hello!
I am new to AWS Glue and I starting to create data monitoring rules in AWS Glue. I have tried multiple options with CustomSQL but can not seem to find the solution.
My problem: I want to check...
1
answers
0
votes
65
views
asked 7 days agolg...
Getting error while connecting streaming data from kinesis to redshift with few transformations using visual ETL. (using amazon kinesis - glue data catalog table as source ). Schema is already...
1
answers
0
votes
131
views
asked 10 days agolg...
We are using Tableau and Tableau has a schedule querying athena.
It worked well until yesterday but I got below issue today.
> HIVE_CANNOT_OPEN_SPLIT: Error opening Hive split...
1
answers
0
votes
259
views
asked 15 days agolg...
Hello,
I have an AWS Glue job that is only supposed to perform an SQL query on the current status. Unfortunately, I always get the following error: "Error Category: QUERY_ERROR; AnalysisException:...
1
answers
0
votes
232
views
asked 17 days agolg...
Question:
We currently have approximately 100 tables in delta format, partitioned by yyyy, mm, dd, hh, mm. Our current process involves reading these delta tables via a crawler, cataloging them, and...
0
answers
0
votes
344
views
asked 18 days agolg...
Reading few gb say 15gb of parquet skewed data , after few transformation such as data type change for some columns and then doing repartitions (dataframe.repartition(120)) before writing it to s3 in...
1
answers
0
votes
258
views
asked 20 days agolg...
I have a glue job which pushes the data from glue into open search.
The index Id column is automatically created while inserting the data into open search.
I would like to pass the index id _id...
1
answers
0
votes
244
views
asked 23 days agolg...
How can I "automatically" add new partitions to a Glue table based on a Hive formatted S3 bucket?lg...
I have a Bucket containing AWS AppStream logs on format `s3://appstream-logs.../sessions/schedule=DAILY/year=2024/month=04/day=03/daily-session-report-2024-04-03.csv`. I have made this data available...
2
answers
0
votes
105
views
asked 23 days agolg...
I am working on migrating data from MySQL to S3 using AWS DMS. I want to employ wildcard mapping for the schema name in the DMS task's selection rules. Specifically, I aim to include tables from...
6
answers
0
votes
163
views
asked 25 days agolg...