A little question here.
Making a code in CUDA, I needed a variable shared memory size. I believed I couldn’t do that, but the function cudaConfigureCall then dazzled me.
I don’t quite know how to work with that function, specifically how to setup the shared memory size.
So my question is, can I use cudaConfigureCall to call a kernel with a variable-depending shared memory size, instead of a constant shared memory size?