Questions tagged with Amazon EC2
Content language: English
Sort by most recent
I'm trying to install the CUDA toolkit to a g4dn.xlarge EC2 instance (Debian Linux 11) but it ran out of memory. The root volume is only 8GB but the instance volume is 125 GB.
Here's the commands that I'm using to get and install the [CUDA toolkit](https://developer.nvidia.com/cuda-downloads?target_os=Linux&target_arch=x86_64&Distribution=Debian&target_version=11&target_type=deb_local):
```
wget https://developer.download.nvidia.com/compute/cuda/12.1.0/local_installers/cuda-repo-debian11-12-1-local_12.1.0-530.30.02-1_amd64.deb
sudo dpkg -i cuda-repo-debian11-12-1-local_12.1.0-530.30.02-1_amd64.deb
```
I also tried using the dpkg option --root=/instance after mounting the instance volume to /instance which fixed the out-of-memory error but led to another error since it expects to be installed at / not at /instance
Can anyone help me get the CUDA toolkit installed please?
I am having issues in regards to adding storage onto my instances I have already created and have been using for awhile. My storage is low so I decided to add more storage. I added 30GB of storage to each instance (General Purpose SSD (gp2) changed to 60 GiB from 30 GiB, when I sign onto the server it doesn’t show up in the storage but it shows up in my disk management that I have 30GB unallocated. What do i need to do to get the 30GB I added to the instance onto the server itself?
Hello every body, i subscribed for Docker Engine - Enterprise for Windows Server 2019 and followed this video
https://www.youtube.com/watch?v=7eObt3MSzWw&ab_channel=CloudInfrastructureServices
but i encountered the the below error :
Success Restart Needed Exit Code Feature Result
------- -------------- --------- --------------
False Maybe Failed {}
Install-WindowsFeature : A prerequisite check for the Hyper-V feature failed.
1. Hyper-V cannot be installed: The processor does not have required virtualization capabilities.
Please advise.
i have attached private subnet 1c to public application load balancer. what would happen with the server which is in public subnet 1c. will the traffic reach to that server ?
I follow this blog to try the hudi connect: [Ingest streaming data to Apache Hudi tables using AWS Glue and Apache Hudi DeltaStreamer](https://aws.amazon.com/cn/blogs/big-data/ingest-streaming-data-to-apache-hudi-tables-using-aws-glue-and-apache-hudi-deltastreamer/).
But when I started the glue job, I always got this error log:
```
2023-03-28 12:39:33,136 - __main__ - INFO - Glue ETL Marketplace - Preparing layer url and gz file path to store layer 8de5b65bd171294b1e04e0df439f4ea11ce923b642eddf3b3d76d297bfd2670c.
2023-03-28 12:39:33,136 - __main__ - INFO - Glue ETL Marketplace - Getting the layer file 8de5b65bd171294b1e04e0df439f4ea11ce923b642eddf3b3d76d297bfd2670c and store it as gz.
Traceback (most recent call last):
File "/usr/lib64/python3.7/runpy.py", line 193, in _run_module_as_main
"__main__", mod_spec)
File "/usr/lib64/python3.7/runpy.py", line 85, in _run_code
exec(code, run_globals)
File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 361, in <module>
main()
File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 351, in main
res += download_jars_per_connection(conn, region, endpoint, proxy)
File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 304, in download_jars_per_connection
download_and_unpack_docker_layer(ecr_url, layer["digest"], dir_prefix, http_header)
File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 168, in download_and_unpack_docker_layer
layer = send_get_request(layer_url, header)
File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 80, in send_get_request
response.raise_for_status()
File "/home/spark/.local/lib/python3.7/site-packages/requests/models.py", line 941, in raise_for_status
raise HTTPError(http_error_msg, response=self)
requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://prod-us-east-1-starport-layer-bucket.s3.us-east-1.amazonaws.com/6a636e-709825985650-a6bdf6d5-eba8-e643-536c-26147c8be5f0/84e9f346-bf80-4532-ac33-b00f5dbfa546?X-Amz-Security-Token=....Ks4HlEAQcC0PUIFipDGrNhcEAVTZQ%3D%3D&X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Date=20230328T123933Z&X-Amz-SignedHeaders=host&X-Amz-Expires=3600&X-Amz-Credential=%2F20230328%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Signature=c28f35ab3b3c
Glue ETL Marketplace - failed to download connector, activation script exited with code 1
LAUNCH ERROR | Glue ETL Marketplace - failed to download connector.Please refer logs for details.
Exception in thread "main"
java.lang.Exception: Glue ETL Marketplace - failed to download connector.
at com.amazonaws.services.glue.PrepareLaunch.downloadConnectorJar(PrepareLaunch.scala:1043)
at com.amazonaws.services.glue.PrepareLaunch.com$amazonaws$services$glue$PrepareLaunch$$prepareCmd(PrepareLaunch.scala:759)
at com.amazonaws.services.glue.PrepareLaunch$.main(PrepareLaunch.scala:42)
at com.amazonaws.services.glue.PrepareLaunch.main(PrepareLaunch.scala)
```
I guess the root cause is:
1. The Glue job cannot pull the connect image from AWS maketplace.
2. The connector image cannot store into the S3 bucket.
So I try these methods:
1. Give permissions to the IAM role of the job. I give `AWSMarketplaceFullAccess, AmazonEC2ContainerRegistryFullAccess, AmazonS3FullAccess`, I think these permissions are enough definitely.
2. Make the S3 bucket public. I turned off the `Block public access` of the related S3 bucket.
But even I did these, I still got the same error. Can someone give any suggestions?
Hello, I would like to expand a partitation on an instance, but for some reason the command growpart does not exist on this instance?
I am proceeding as usual according to https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/recognize-expanded-volume-linux.html, and this has also worked for the other instances in the past, only with the current one I fail at point 2c, as the command "growpart" obviously does not exist on my instance?!?
This is the current situation at this instance:
-------------
sudo lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
xvda 202:0 0 8G 0 disk
`-xvda1 202:1 0 5G 0 part /
xvdb 202:16 0 15G 0 disk
`-xvdb1 202:17 0 15G 0 part /IOL
-------------
So the xvda drive has already been increased from 5 to 8 GB.
But when I try to enlarge it using growpart, the result looks like this:
-----------
sudo growpart /dev/xvda 1
sudo: growpart: command not found
-------
And indeed, there is obviously none:
--------
which growpart
/usr/bin/which: no growpart in (/usr/lib64/qt-3.3/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/ec2-user/bin)
----------
Maybe you can imagine the look on my face when I saw this message? 8:o
Where did that go?! Are there alternative ways to increase the partitation to the full drive size without growpart?
I received an email from Amazon and was informed that the instance scheduled for retirement. After I started the instance, I found that the network interface, security group, and elastic IP information associated with the instance could not be found.(The instance was officially stop by Amazon)
How can I find my information back.
Hello,
My client's EC2 instance has been kept crashing randomly for a while, instance ID: i-0f4d7b0f00f614d62
When it went down, I couldn't ssh into it. I posted a question on serverfault: https://serverfault.com/questions/1127277/ec2-instance-randomly-crash
But some times it doesn't seem to be a memory issue, while I was monitoring memory usage, the system could suddenly go down even when it had sufficient memory at the last minute.
Could you investigate please?
Hello,
I am trying to setup SSO between an EC2 instance running Gitlab and an EC2 instance with Active Directory following this documentation : "https://docs.gitlab.com/ee/ci/cloud_services/aws/".
I have tried to add our Gitlab address in IAM Identity Providers but I have this message "could not connect to openid configuration of provider".
The issue, I think, is that our Active Directory EC2 instance is running under a different account than our Gitlab EC2 instance.
Also, our Gitlab EC2 instance is registered in a private zone in Route 53 , under this different account.
Do you know how we can allow communication between EC2 instances running under different account ?
Rémi
Hi Team,
We are unable to start Instance WE1T105410002 and getting 1/2 Status check failed. We are also unable to check system log as they are empty. Could you please check from your end.
Regards
Saud Salman
Apologies if this is a naive question or not the correct place to asked, new to AWS and not finding answers online.
I am working with sequencing data in r. Based on some research, using AWS EC2 rstudio seemed like a good approach for me since I do not have access to servers at my institution. I uploaded my data files from an external drive to S3. My challenge is how to work with the multiple files in r. From what I understand, you cannot use S3 as a working directory, just as a location to store the files. I also believe r has a memory limitation of 1 GB. I have multiple files that total approximately 14 GB. So reading the files into r is not a solution. Any suggestions around this, other solutions, or tutorials available?
The domain name I applied for in godaddy is bound to AWS Global Accelerator. The endpoint of AWS Global Accelerator adds two ALBs in different regions. The ALB listens to the EC2 of the web server. When both of my web services fail, how should I set it up? Failover to a static maintenance page. The current idea is to add an A record through the R53 regional hosting to failover to the static website hosting of the S3 bucket of the same domain name, but it does not take effect and cannot jump to the S3 static page. How to set up to jump, Or is there a third party hosted maintenance page that fails over in some other way