I’m trying to convert onnx model to tensorrt with batchsize 64.
During inference,it takes 390ms for batch inference (batch_size = 64)
and 7 ms for batch size = 1.
Please let me know if I’m missing out on something.
Attaching google drive link for tensorrt model generation from onnx where build_retina_trt.py where it converts mxnet-> onn → tensorrt.
I was able to generate the model successfully, only thing is during the inference the performance is not what i expected. Hence sharing the github repository. here