Hi, we’d like to run TensorRT 6 or 8 on both Jetson AGX and in AWS (Turing T4 on Arm). Do we have to build different engines locally on each platform or is there a way to generate PTX or at least force TRT to generate multiple CCs from one platform run?
Tensorrt doesn’t support that. Yes, we have to build different engines locally on each platform
The generated plan/engine files are not portable across platforms or TensorRT versions. Plans are specific to the exact GPU model they were built on (in addition to the platforms and the TensorRT version) and must be re-targeted to the specific GPU.
Thank you for confirming. You guys may want to add support for PTX/alternate path/or cuDNN fallback for use cases when performance is not important but let’s say the same code needs to run in CI pipeline on a card with a different CC.
Because the cost of switching versions/different CC is sometimes stratospheric. I spent the last 120 hrs trying to get running on CC 8.6 the model that was already previously running and exporting fine on Turing. Can’t say i’m a huge fan of TensorRT at the moment.