I have build a pipeline to export a EfficientDet Pytorch implementation first to ONNX and then from ONNX to TensorRT with ONNX-Tensorrt. The conversion seems to work alright. I only get the following warning when exporting to TensorRT which shouldn’t be the cause of the problem:
Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
But when doing inference with TensorRT the output is always zero while in ONNX and in PyTorch I get a non zero output for the same input.
I also tested my TensorRT inference script with a minmal example model and it got a non zero output with the same pipeline. Am I missing something why am I only getting zero outputs?
Environment
**TensorRT Version 7.1.3.0:
**GPU Type Nvidia Jetson Xavier:
**Nvidia Driver Version N/A:
**CUDA Version 10.2:
**CUDNN Version N/A:
**Operating System + Version Ubuntu 18.04:
**Python Version (if applicable) python 3.7:
**PyTorch Version (if applicable) 1.7:
For the TensorRT conversion I used: onnx-tensorrt
Use the onnx inference script to run successful inference on the sample image
Use the TRT script to run non successful inference on the the sample image
TensorRT does not natively support INT64, whereas in Onnx, some operators require int_max or int_min as special values to denote ‘infinity’(e.g. Slice operator), which is probably where the large integer values are coming from.