Hello everyone!
Does anybody know how to force double precision computation in CUDA programs. Command line contains information about GPU compatibility:
"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v3.2\bin\nvcc.exe" -gencode=arch=compute_13,code=\"sm_13,compute_13\" --use-local-env --cl-version 2008 -ccbin "g:\DevTools\Microsoft Visual Studio 9.0\VC\bin\x86_amd64" -I"C:\Program Files\NVIDIA GPU
Computing Toolkit\CUDA\v4.0\lib\x64" -I"../../common/lib" -I"C:\ProgramData\NVIDIA Corporation\NVIDIA GPU Computing SDK 3.2\C\common\inc\dynlink" -I"C:\Program Files (x86)\NVIDIA GPU Computing Toolkit\CUDA\v3.2\include"
-I"C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v3.2\include" -G0 --keep-dir "x64\Debug\" -maxrregcount=32 --machine 64 --compile -D_NEXUS_DEBUG -g -Xcompiler "/EHsc /nologo /Od /Zi /MDd " -o "x64\Debug\MatMul.obj"
"C:\Users\insomnia\Documents\Visual Studio 2010\Projects\CUDATest\CUDATest\MatMul.cu"
But execution performance does not differ in case of using single and double precision variables in code. Does anyone know how to force DP ?