on Titan RTX and also Turing-family GeForce GPUs such as 2080/2080Ti, P2P is only supported if/when the NVLink bridge is in place (i.e. only over NVLink). For Turing family GeForce GPUs without a NVLink bridge option, P2P is not supported.
What NVIDIA supports is determined by the tool already run. The behavior may vary by GPU, and is not strictly a function of the motherboard. Since I am not providing a matrix that covers every question for every case, the best determinant of features like this is what the tool reports.
NVIDIA may choose to design a product that supports P2P over NVLink but not PCIE. That is what happened with Titan RTX. I won’t be able to give further rationalization or explanation.
In particular, questions like “why is it this way?” I am not able to answer.
For example, “Anyone knows what makes this difference ?” is a question I wouldn’t be able to answer, other than by saying that is the way the product is designed.
I have two Titan RTXs with an NVlink between them. From what I can see, each card can accommodate just one NVlink so you would end up with pairs of them connected. As it turns out, I doubt I will need any connection between them for my purposes but it’s nice to have in case we do need it.
Yeah, mebe it’s just worth restating what the impact of p2p memory copy is? I’m interested in tensorflow and pytorch deep learning over multiple GPUs. Atm we are 1080tis using pcie, but soon upgrading to Turing/RTX cards.
Any info much appreciated as atm it doesn’t look like it’s worth upgrading, but it could be just a poor appreciation of the impact of p2p.
On what level this behavior is defined?
If its part of the driver we (external developers) may be able to change it, if its part of internal architecture - we probably won’t.
Not everyone needs NVlinks, its a product to increase communication BW, but BW is not always the bottleneck.
When we don’t want to use IB, or when BW is not the bottleneck, NVlinks improve performance sometimes only by 1%, therefore may not worth the price.
By doing this design NVidia coupled Turing-family GeForce GPUs to another (sometimes unnecessary) product.
I wonder which HW should we buy if we want to use use P2P with “commodity” communication.
This appears to be an intentional “gimp” to push professional users to quadro / tesla professional cards for memory pooling. Nvidia, probably correctly, reasons that without such “gimp” commercial users would be tempted to use the RTX Titans in a server type environment despite the warnings not to.