[b][font=“'Lucida Console”]I have several independent Host processes that all have GPU kernels they want to run on their own data. I am trying to understand [/font]
[font=“'Lucida Console”]how a CUDA GPU board gets shared among the Host processes. I am guessing that each Host process inaugurates its own context and thus each host process can use the entire global memory on the board.[/font]
[font=“'Lucida Console”]I am trying to understand the blocking and sharing mechanisms. specifically if the GPU is busy executing a kernel for one host process and a second host process tries to access the gpu board then does the[/font]
[font=“'Lucida Console”]second process block until the GPU is free? [/font]
[font=“'Lucida Console”]
[/font]
[font=“'Lucida Console”]I would like to hear from Nvidia engineers who can reference specific pages in Nvidia documentation that describes this phenomena.[/font]
[font=“'Lucida Console”]
[/font]
[font=“'Lucida Console”]thanks[/font]
[font=“'Lucida Console”]
[/font]
[/b]