Hi all,
To reproduce the NVIDIA Jetson AGX Xavier benchmarks of Tiny YOLO V3, i’ve used the sample /usr/src/tensorrt/samples/python/yolov3_onnx but i’ve got perfs between 600 and 700 FPS with this parameters link of my recent post which contains thr details of the implemntation:
Power Mode : MAXN
Input resolution : 416x416
Precision Mode : INT8 (Calibration with 1000 images and IInt8EntropyCalibrator2 interface)
batch = 8
JetPack Version : 4.5.1
TensorRT version : 7.1.3
I’ve also run this command to maximize the perfs
$ sudo nvpmodel -m 0
$ sudo jetson_clocks
Then i tried GitHub - NVIDIA-AI-IOT/jetson_benchmarks: Jetson Benchmark to reproduce the FPS of Tiny YOLO V3 and that worked properly (i got the 1000 FPS), but i noticed that this repos use the shell trtexec and not the python API tensorrt that i’am used to use by (import tensorrt as trt)
My question is what’s the difference between this two implementations ( trtexec VS tensorrt python API) when we generate/build the TensorRT engine and when we run the inference ?
Thanks