I am trying to run a CUDA piece of code which is a simple Hello World program on the NVIDIA Tesla. I know for certain that the NVIDIA Tesla device is connected as when I try /bin/lspci it shows a long list of connected devices ending with the Tesla units which ends with the code seen below.
30:00.0 3D controller: nVidia Corporation G80 [Tesla C870] (rev a2)
However when I try running the basic Hello world program that is supposed to mangle the hello world in the device(GPU) and then print the mangled gpu code on the host, it never gets mangled. The hello world that I am trying to get working is the first example found on: [url=“The Official NVIDIA Forums | NVIDIA”]The Official NVIDIA Forums | NVIDIA
I know that the code for the Hello World is correct.
So in short, even though when I do nvcc blah.cu there is no error; when I try and execute ./a.out the output is never mangled, implying that the gpu code is never executed Only the CPU code is functional
Am I supposed to be specifying the port number to which the Tesla is connected or does the computer automatically know it? if so how could ispecify the port number? All the examples that I see online seem to assume that the system knows which port the GPU is connected to.
Also wrapper functions such as CUDA_SAFE_CALL do not seem to work. Is the because wrapper functions need a special header file to be included in the code?
When I tried the command nvidia-settings to figure out the driver version i get
nvidia-settings: error while loading shared libraries: libgtk-x11-2.0.so.0: cannot open shared object file: No such file or directory
I think that may be because the sdk is not installed
When I tried the command nvidia-settings to figure out the driver version i get
nvidia-settings: error while loading shared libraries: libgtk-x11-2.0.so.0: cannot open shared object file: No such file or directory
I think that may be because the sdk is not installed
That looks ok. If you have /dev/nividia* entries with appropriate permissions it should work. At this point it might be worth getting the SDK and try building and running the deviceQuery sample.
That looks ok. If you have /dev/nividia* entries with appropriate permissions it should work. At this point it might be worth getting the SDK and try building and running the deviceQuery sample.
If you don’t have those, something is wrong. Are you running X11? If not, you need to (re)read the toolkit release notes. There is information about setting up an init script which will create those device file entries at boot time.
If you don’t have those, something is wrong. Are you running X11? If not, you need to (re)read the toolkit release notes. There is information about setting up an init script which will create those device file entries at boot time.
Thanks, that helped a lot to identify the problem. The instructions that I followed on adding the script to the .profile file in order to initiate x11 automatically didn’t work. I wonder why CUDA never shows an error when i do nvcc blah.cu and then ./a.out. It just doesn’t do anything in the CUDA device. Is there an easier way to launch the application without admin rights? Admin rights take a while to get approved.
Thanks, that helped a lot to identify the problem. The instructions that I followed on adding the script to the .profile file in order to initiate x11 automatically didn’t work. I wonder why CUDA never shows an error when i do nvcc blah.cu and then ./a.out. It just doesn’t do anything in the CUDA device. Is there an easier way to launch the application without admin rights? Admin rights take a while to get approved.