Hi, Friend!
I am a beginer for CUDA.
As I know, a warp is the number of threads to be executed concurrently by a multiprocessor.
-
If it is right, does foo<<<100, 32> mean that foo will be ececuted 100 times by a warp?
If it is right, it means that foo should be executed by only a multiprocessor even if my graphic card has 4 multiprocessors. -
If my card has 4 multiprocessors, does foo<<<100, 32>>> mean that foo will be executed 25 times by each multiprocessor?
-
If my card has 4 multiprocessors, does foo<<<100, 16>>> also mean that foo will be executed 25 times by each multiprocessor?
-
If my card has 4 multiprocessors, does foo<<<1, 90>>> mean that foo will be executed once by 3 multiprocessors?
-
In the 4th case, can 90 threads use the same shared memory? I know that the threads in a multiprocessor only can access the same shared emnory.
Please be generous for my poor English.