Nvidia says that the new Pascal P100 cards improves the overlapping of workloads with Asynchronous Concurrent Compute.
But, is it truly possible to share the GPU with different workloads in the P100?
I’ve read in the NVIDIA manual for Kepler GPUs that “The GPU has a time sliced scheduler to schedule work from work queues belonging to different CUDA contexts. However, work launched to the compute engine from work queues belonging to different CUDA contexts cannot execute concurrently.”
Is it possible to execute different CUDA contexts concurrently in P100?