Tensorrt 4 for Jetson Tx1 with cuda8.0

Dear all,

I am currently seeking a TensorRT 4 build compatible with TX1 and CUDA 8.0, specifically for Jetpack 28.1. While I’ve located a build for amd64, the same is not available for arm64 (Jetson).

My goal is to employ TensorRT 4 to “compile” a PyTorch model for inference purposes. It’s crucial to note that I cannot perform a physical flash on the TX1 units, as I manage over 100 devices globally, updating them exclusively through scripts.

I am working with PyTorch version 1.1.0 (though I’m uncertain if the version is a critical factor). When using the PyTorch module directly with “[0]” on the CUDA device, it consumes more than 1.5 GB of memory.

If anyone has managed to acquire the TensorRT 4 build for TX1 or has alternative solutions, your insights would be greatly appreciated.

Best regards,

Manel

Hi,

Have you tried to compile the model with TensorRT 4 on your AMD environment?
Since the TensorRT 4 was released years ago, it might not be able to support the recent model.

Thanks.