1 Answer
- Newest
- Most votes
- Most comments
0
If you are using a Sagemaker notebook attached to Glue dev endpoint with PySpark kernel, please use the following method to update the libraries.
- Update the dev endpoint using
aws glue update-dev-endpoint --cli-input-json file://./input.json
where input.json looks like
{
"EndpointName": "YOUR_DEV_ENDPOINT_NAME",
"CustomLibraries": {
"ExtraPythonLibsS3Path": "s3://my-bucket/library1.zip,s3://my-bucket/library2.zip,s3://my-bucket/library3.zip"
},
"UpdateEtlLibraries": true
}
- You can then verify the new list of libraries using following Python code:
import os
os.listdir('.')
Please note that as of writing this, Sagemaker Jupyter notebooks running PySpark kernel now come with following Python version
3.6.12 (default, May 18 2021, 22:47:55) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)]
Relevant content
- asked 3 years ago
- AWS OFFICIALUpdated 5 months ago
- AWS OFFICIALUpdated 2 years ago
- How do I install and troubleshoot Python libraries in Amazon EMR and Amazon EMR Serverless clusters?AWS OFFICIALUpdated 7 months ago
- AWS OFFICIALUpdated 8 months ago