I’m converting a TensorFlow graph into TensorRT engine. The targeted device for deployment is 1080 Ti. However my desk machine has only 1080. Given that both devices have compute capability 6.1, I wonder if I can optimize TensorRT engine on 1080 while expecting getting optimized performance when deployed on 1080Ti.
We created a new “Deep Learning Training and Inference” section in Devtalk to improve the experience for deep learning and accelerated computing, and HPC users:
We are moving active deep learning threads to the new section.
URLs for topics will not change with the re-categorization. So your bookmarks and links will continue to work as earlier.
There is no guarantee that the performance characteristics of a network when optimized for one device will be the same on another device. That being said, the two devices are similar enough that it is unlikely that performance will diverge. The only way to be sure is to test. If there is a functional issue, please file a bug here: https://developer.nvidia.com/nvidia-developer-program
Please include the steps/files used to reproduce the problem along with the output of infer_device.
@mvillmow - I have seen you post the infer_device tool in many of these threads, and I can’t seem to find this tool, does it even exist?
Sorry, infer_device is the name of the deviceQuery sample that we build for inference. Please use deviceQuery from the cuda samples to get the information required. We don’t ship infer_device because it is the same source code.