1 Answer
- Newest
- Most votes
- Most comments
0
【以下的回答经过翻译处理】 你可以在Spark作业中使用Hadoop Java配置来更新S3 canned acl。
‘’‘
import sys from pyspark import SparkConf from pyspark.context import SparkContext from pyspark.sql.functions import * from pyspark.sql import SQLContext, Row
sc = SparkContext()
’‘’
Relevant content
- asked 10 months ago
- Accepted Answerasked 2 months ago
- asked a year ago
- asked a year ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated 5 months ago
- AWS OFFICIALUpdated a year ago
- AWS OFFICIALUpdated a year ago