nvidia-smi not working on Amazon's DeepLearning AMI

0

Greetings,

I have set up an instance of "Deep Learning AMI (Ubuntu)" (https://aws.amazon.com/marketplace/pp/B077GCH38C)
expecting it to have functioning NVIDIA GPUs.

However, when I write nvidia-smi in the command line ,I get:

NVIDIA-SMI has failed because it couldn't communicate with the NVIDIA driver. Make sure that the latest NVIDIA driver is installed and running.

And when I try to execute Python code that uses PyTorch 0.4 (I did install that version as required) I get:

THCudaCheck FAIL file=/opt/conda/conda-bld/pytorch_1532579245307/work/aten/src/THC/THCGeneral.cpp line=74 error=38 : no CUDA-capable device is detected
Traceback (most recent call last):
  File "main.py", line 105, in <module>
    train_data = batchify(corpus.train, args.batch_size, args)
  File "/home/ubuntu/awd-lstm-lm/utils.py", line 21, in batchify
    data = data.cuda()
RuntimeError: cuda runtime error (38) : no CUDA-capable device is detected at /opt/conda/conda-bld/pytorch_1532579245307/work/aten/src/THC/THCGeneral.cpp:74

Has anyone experienced something similar? Any way to fix this?

AndreaL
已提問 5 年前檢視次數 1250 次
1 個回答
0

Answered: I did not choose a GPU instance when setting up the machine

AndreaL
已回答 5 年前

您尚未登入。 登入 去張貼答案。

一個好的回答可以清楚地回答問題並提供建設性的意見回饋,同時有助於提問者的專業成長。

回答問題指南