We currently have CUDA software running on a single Tesla GPU. However we plan to subdivide the problem and run it on multiple GPUs.
Since considerable data transfer between the GPUs might be needed, we want to make use of CUDA 4.0 capabilities for multi GPU programming.
Apart from speed, also the software maintanance and transparent programming of multiple GPUs is crucial.
For a PC with 2 GPUs we don not see a problem.
However we have now been offered the possibility to configure/buy a system supporting upto 8 Tesla GPUs.
As far as I understand (I am by no means a hardware expert) the system basicall consists of multiple mainboards which are intelligently coupled and I know that such multi-processor systems are around.
Does anyone have experience with such systems, possibly with using multiple GPUs on them ?
Can I expect/hope that CUDA 4.0 will support more than 2 GPUs on such a system transparently (GPU-GPU data transfer, global address space etc.) ?
Or are there pitfalls/limitations to be expected in principle ?
kind regards Rolf