Is there any documentation on the amount of memory that is occupied by allocating a cudaArray as a function of type/width/height/depth? If not, does there exist a function that returns the result (even without explaining how it got there)? Does the result depend on the used cuda driver or the GPU?
I often need to allocate large cudaArray:s and would like to know in advance whether they will fit in the free memory. I have been experimenting for a while … for a float-cudaArray the size seems to be at least rounded up to the next multiple of 2^19 elements, but sometimes it’s more. I can also see, that width/height/depth are not interchangeable.
Out of curiosity, is there any official documentation of the internal layout of a cudaArray (clearly I’m not indexing into it or such…)? There are people speculating about the use of a spacefilling curve, but I don’t see any official statements about that.
cheers, Lukas Wirz