Hi, all
I have read the following information from a PPT,
"Contemporary (Fermi) GPU Architecture
32 CUDA Cores per Streaming Multiprocessor (SM)
32 fp32 ops/clock
16 fp64 ops/clock
32 int32 ops/clock
2 Warp schedulers per SM
1,536 concurrent threads
4 special–â€function units
64KB shared memory + L1 cache
32K 32–â€bit registers
Fermi GPUs have as many as 16 SMs
24,576 concurrent threads "
I can’t figure out how 1,536 is obtained.
Even though I have read the whitepaper: NVIDIA’s Next Generation CUDATM Compute Architecture: Fermi.
Is the number of concurrent threads related to the number of warp schedulers?
In my view, the number is determined by the hardware computer resources.