Questions tagged with Amazon EC2

Content language: English

Sort by most recent

Browse through the questions and answers listed below or filter and sort to narrow down your results.

I'm trying to install the CUDA toolkit to a g4dn.xlarge EC2 instance (Debian Linux 11) but it ran out of memory. The root volume is only 8GB but the instance volume is 125 GB. Here's the commands that I'm using to get and install the [CUDA toolkit](https://developer.nvidia.com/cuda-downloads?target_os=Linux&target_arch=x86_64&Distribution=Debian&target_version=11&target_type=deb_local): ``` wget https://developer.download.nvidia.com/compute/cuda/12.1.0/local_installers/cuda-repo-debian11-12-1-local_12.1.0-530.30.02-1_amd64.deb sudo dpkg -i cuda-repo-debian11-12-1-local_12.1.0-530.30.02-1_amd64.deb ``` I also tried using the dpkg option --root=/instance after mounting the instance volume to /instance which fixed the out-of-memory error but led to another error since it expects to be installed at / not at /instance Can anyone help me get the CUDA toolkit installed please?
0
answers
0
votes
5
views
Jasper
asked an hour ago
I am having issues in regards to adding storage onto my instances I have already created and have been using for awhile. My storage is low so I decided to add more storage. I added 30GB of storage to each instance (General Purpose SSD (gp2) changed to 60 GiB from 30 GiB, when I sign onto the server it doesn’t show up in the storage but it shows up in my disk management that I have 30GB unallocated. What do i need to do to get the 30GB I added to the instance onto the server itself?
2
answers
0
votes
14
views
asked 2 hours ago
Hello every body, i subscribed for Docker Engine - Enterprise for Windows Server 2019 and followed this video https://www.youtube.com/watch?v=7eObt3MSzWw&ab_channel=CloudInfrastructureServices but i encountered the the below error : Success Restart Needed Exit Code Feature Result ------- -------------- --------- -------------- False Maybe Failed {} Install-WindowsFeature : A prerequisite check for the Hyper-V feature failed. 1. Hyper-V cannot be installed: The processor does not have required virtualization capabilities. Please advise.
1
answers
0
votes
5
views
asked 3 hours ago
i have attached private subnet 1c to public application load balancer. what would happen with the server which is in public subnet 1c. will the traffic reach to that server ?
1
answers
0
votes
9
views
Cfr
asked 5 hours ago
I follow this blog to try the hudi connect: [Ingest streaming data to Apache Hudi tables using AWS Glue and Apache Hudi DeltaStreamer](https://aws.amazon.com/cn/blogs/big-data/ingest-streaming-data-to-apache-hudi-tables-using-aws-glue-and-apache-hudi-deltastreamer/). But when I started the glue job, I always got this error log: ``` 2023-03-28 12:39:33,136 - __main__ - INFO - Glue ETL Marketplace - Preparing layer url and gz file path to store layer 8de5b65bd171294b1e04e0df439f4ea11ce923b642eddf3b3d76d297bfd2670c. 2023-03-28 12:39:33,136 - __main__ - INFO - Glue ETL Marketplace - Getting the layer file 8de5b65bd171294b1e04e0df439f4ea11ce923b642eddf3b3d76d297bfd2670c and store it as gz. Traceback (most recent call last): File "/usr/lib64/python3.7/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/usr/lib64/python3.7/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 361, in <module> main() File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 351, in main res += download_jars_per_connection(conn, region, endpoint, proxy) File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 304, in download_jars_per_connection download_and_unpack_docker_layer(ecr_url, layer["digest"], dir_prefix, http_header) File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 168, in download_and_unpack_docker_layer layer = send_get_request(layer_url, header) File "/tmp/aws_glue_custom_connector_python/docker/unpack_docker_image.py", line 80, in send_get_request response.raise_for_status() File "/home/spark/.local/lib/python3.7/site-packages/requests/models.py", line 941, in raise_for_status raise HTTPError(http_error_msg, response=self) requests.exceptions.HTTPError: 403 Client Error: Forbidden for url: https://prod-us-east-1-starport-layer-bucket.s3.us-east-1.amazonaws.com/6a636e-709825985650-a6bdf6d5-eba8-e643-536c-26147c8be5f0/84e9f346-bf80-4532-ac33-b00f5dbfa546?X-Amz-Security-Token=....Ks4HlEAQcC0PUIFipDGrNhcEAVTZQ%3D%3D&X-Amz-Algorithm=AWS4-HMAC-SHA256&X-Amz-Date=20230328T123933Z&X-Amz-SignedHeaders=host&X-Amz-Expires=3600&X-Amz-Credential=%2F20230328%2Fus-east-1%2Fs3%2Faws4_request&X-Amz-Signature=c28f35ab3b3c Glue ETL Marketplace - failed to download connector, activation script exited with code 1 LAUNCH ERROR | Glue ETL Marketplace - failed to download connector.Please refer logs for details. Exception in thread "main" java.lang.Exception: Glue ETL Marketplace - failed to download connector. at com.amazonaws.services.glue.PrepareLaunch.downloadConnectorJar(PrepareLaunch.scala:1043) at com.amazonaws.services.glue.PrepareLaunch.com$amazonaws$services$glue$PrepareLaunch$$prepareCmd(PrepareLaunch.scala:759) at com.amazonaws.services.glue.PrepareLaunch$.main(PrepareLaunch.scala:42) at com.amazonaws.services.glue.PrepareLaunch.main(PrepareLaunch.scala) ``` I guess the root cause is: 1. The Glue job cannot pull the connect image from AWS maketplace. 2. The connector image cannot store into the S3 bucket. So I try these methods: 1. Give permissions to the IAM role of the job. I give `AWSMarketplaceFullAccess, AmazonEC2ContainerRegistryFullAccess, AmazonS3FullAccess`, I think these permissions are enough definitely. 2. Make the S3 bucket public. I turned off the `Block public access` of the related S3 bucket. But even I did these, I still got the same error. Can someone give any suggestions?
1
answers
0
votes
5
views
donglai
asked 6 hours ago
Hello, I would like to expand a partitation on an instance, but for some reason the command growpart does not exist on this instance? I am proceeding as usual according to https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/recognize-expanded-volume-linux.html, and this has also worked for the other instances in the past, only with the current one I fail at point 2c, as the command "growpart" obviously does not exist on my instance?!? This is the current situation at this instance: ------------- sudo lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT xvda 202:0 0 8G 0 disk `-xvda1 202:1 0 5G 0 part / xvdb 202:16 0 15G 0 disk `-xvdb1 202:17 0 15G 0 part /IOL ------------- So the xvda drive has already been increased from 5 to 8 GB. But when I try to enlarge it using growpart, the result looks like this: ----------- sudo growpart /dev/xvda 1 sudo: growpart: command not found ------- And indeed, there is obviously none: -------- which growpart /usr/bin/which: no growpart in (/usr/lib64/qt-3.3/bin:/usr/local/bin:/bin:/usr/bin:/usr/local/sbin:/usr/sbin:/sbin:/home/ec2-user/bin) ---------- Maybe you can imagine the look on my face when I saw this message? 8:o Where did that go?! Are there alternative ways to increase the partitation to the full drive size without growpart?
0
answers
0
votes
1
views
asked 8 hours ago
I received an email from Amazon and was informed that the instance scheduled for retirement. After I started the instance, I found that the network interface, security group, and elastic IP information associated with the instance could not be found.(The instance was officially stop by Amazon) How can I find my information back.
0
answers
0
votes
15
views
asked 11 hours ago
Hello, My client's EC2 instance has been kept crashing randomly for a while, instance ID: i-0f4d7b0f00f614d62 When it went down, I couldn't ssh into it. I posted a question on serverfault: https://serverfault.com/questions/1127277/ec2-instance-randomly-crash But some times it doesn't seem to be a memory issue, while I was monitoring memory usage, the system could suddenly go down even when it had sufficient memory at the last minute. Could you investigate please?
1
answers
0
votes
15
views
asked 11 hours ago
Hello, I am trying to setup SSO between an EC2 instance running Gitlab and an EC2 instance with Active Directory following this documentation : "https://docs.gitlab.com/ee/ci/cloud_services/aws/". I have tried to add our Gitlab address in IAM Identity Providers but I have this message "could not connect to openid configuration of provider". The issue, I think, is that our Active Directory EC2 instance is running under a different account than our Gitlab EC2 instance. Also, our Gitlab EC2 instance is registered in a private zone in Route 53 , under this different account. Do you know how we can allow communication between EC2 instances running under different account ? Rémi
1
answers
0
votes
21
views
asked 11 hours ago
Hi Team, We are unable to start Instance WE1T105410002 and getting 1/2 Status check failed. We are also unable to check system log as they are empty. Could you please check from your end. Regards Saud Salman
0
answers
0
votes
2
views
asked 15 hours ago
Apologies if this is a naive question or not the correct place to asked, new to AWS and not finding answers online. I am working with sequencing data in r. Based on some research, using AWS EC2 rstudio seemed like a good approach for me since I do not have access to servers at my institution. I uploaded my data files from an external drive to S3. My challenge is how to work with the multiple files in r. From what I understand, you cannot use S3 as a working directory, just as a location to store the files. I also believe r has a memory limitation of 1 GB. I have multiple files that total approximately 14 GB. So reading the files into r is not a solution. Any suggestions around this, other solutions, or tutorials available?
0
answers
0
votes
3
views
asked 16 hours ago
The domain name I applied for in godaddy is bound to AWS Global Accelerator. The endpoint of AWS Global Accelerator adds two ALBs in different regions. The ALB listens to the EC2 of the web server. When both of my web services fail, how should I set it up? Failover to a static maintenance page. The current idea is to add an A record through the R53 regional hosting to failover to the static website hosting of the S3 bucket of the same domain name, but it does not take effect and cannot jump to the S3 static page. How to set up to jump, Or is there a third party hosted maintenance page that fails over in some other way
0
answers
0
votes
7
views
asked 16 hours ago