Adding the filename as a column in the output

0

Does anyone know of a way to add the source filename as a column in a Glue project? We just started using it, and have it crawling for files in S3, applying some simple transformations, and then writing to a clean file in S3. In our data set we would like to know the source file that each row of data originated from, so it would be ideal if we could somehow add in the filename as a column or something like that. I looked through the documentation and the aws-glue-libs source, but didn't see anything.

I'm still learning Glue, so apologies if I'm using the wrong terminology.

mwatson
已提問 6 年前檢視次數 2818 次
2 個答案
0

You could create folders within your data path and the Glue crawler will create partitions in the schema for every folder. If you name the folders like key=value the column name for the partition will be key.

For instance if your filenames are named source_a..., source_b...
you should create the following

s3/your-bucket/data/source=A/source_a_...csv
s3/your-bucket/data/source=A/source_a_...csv
s3/your-bucket/data/source=B/source_b_...csv
s3/your-bucket/data/source=B/source_b_...csv

so the schema will have a source column for the source partition.

已回答 4 年前
0

Apache Spark has input_file_name() method
You can add a new column based on this. See the below sample.

df = spark.read.parquet('path_to_file')\
        .withColumn('filepath', input_file_name())

See details here: https://spark.apache.org/docs/latest/api/python/pyspark.sql.html?highlight=functions#pyspark.sql.functions.input_file_name

AWS
已回答 4 年前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南