I have two Quadro 5600 in my machine so that i can try some thing with multiple GPU. I have some questions concerning about that
I have a program that run perfectly with one card, it includes multiple kernels, and use some kind of global memory like texture and constant, that normally i define:
texture<float, 2> tex;
device constant int d_extInfo[16]
…
som_kernel()
…
-
In my application i want to run program separately on each CUDA card i have, that mean tex and d_extInfo should be defined locally for each device. How can I do that
-
There’s a common input of the two, how can i allocate in CUDA that so both GPU can access it. Or i have to make two copies of that common input
-
After run program on each GPU, I want to combine the results, in that case i should save the result some where that both GPU can read. Can i store results in GPU, or should i store in CPU and combine results in CPU. I would like to store in GPU so that i can exploit my fast GPU combine function
-
Is there any multi-gpu program that show how to transfer data between GPUs. The multigpu sample in SDK is too simple that i don’t find it good enough to understand what happen
Thank you