I’m new to these forums, but am struggling with an issue and could use getting pointed in the right direction.
I am using a TX1 to perform some fairly memory intensive CUDA calculations.
I’ve noticed that in the CUDA7 samples deviceQuery reports that the GPU only has visibility to 2GB of RAM (plenty for my needs). The OS reports 4GB available, which is great. The trouble that I’m having is that when I try to request memory for my CUDA task, it appears that all the GPU can allocate is the Free memory minus the Kernel Cache. This usually is somewhere around 100 to 200 MB.
Could someone please help me understand how the memory is structured here, and what I can try to do to ensure that there is still plenty of memory available for the GPU?
I experience a similar issue on the K1, but never pursued it. I am hoping that there is something simple in the os that I can configure to get this working.