How to COPY data into Redshift from S3 using a dynamic file path

0

I have several folders in S3 generated from logs in the format: s3://bucketname/2024/03/28/17 s3://bucketname/2024/03/29/18 s3://bucketname/2024/03/28/19 date part '2024/03/28/17' are generated in UTC year, month, date, and hour

I have tried to construct the S3 folder name using || and concat but Redshift doesn't support those, for example: COPY table From 's3://bucketname/2024/' || to_char(getdate(), 'MM') || '/' || to_char(getdate(), 'DD') || '/' || to_char(getdate(), 'HH') IAM_ROLE 'arnxxxx' DELIMITER '\t'

Is there anyway to construct a dynamic file path for COPY command based on dates?

질문됨 한 달 전296회 조회
1개 답변
1
수락된 답변

The COPY command does not support the use of functions or expressions directly within the FROM clause. This means that you cannot use the CONCAT function directly in the FROM clause of the COPY command.

For you archive this desired functionality, you would need to programmatically construct the S3 file path in your client application or script before passing the complete path as a string to the COPY command. You might do this in a Python script:

import datetime
import boto3

# Get the current date and time
now = datetime.datetime.now()

# Construct the S3 file path
s3_path = 's3://bucketname/{}/{}/{}/{}'.format(
    now.year,
    str(now.month).zfill(2),
    str(now.day).zfill(2),
    str(now.hour).zfill(2)
)

# Example Redshift COPY command
copy_command = f"""
COPY table_name
FROM '{s3_path}'
IAM_ROLE 'arn:aws:iam::<account_id>:role/<role_name>'
DELIMITER '\\t';
"""

print(copy_command)
profile picture
전문가
답변함 한 달 전

로그인하지 않았습니다. 로그인해야 답변을 게시할 수 있습니다.

좋은 답변은 질문에 명확하게 답하고 건설적인 피드백을 제공하며 질문자의 전문적인 성장을 장려합니다.

질문 답변하기에 대한 가이드라인

관련 콘텐츠