Hello ,
I am working on with the trt-yolo-app . Inference time on tx2 with yolov2 -416 is aproximately 57 ms . It is fast but after this inference as far as i understand i need to call
inferNet->decodeDetections(imageIdx, curImage.getImageHeight(),curImage.getImageWidth()); functions to get results. It is changing between 11ms to 18ms . Are there any way to reduce duration of decoding? Does it necessary . I don’t know the logic behind this one.
My average speed for processing image is 80 ms . Is it the limit do you think?
I have overclocked the tx2 with nvpmodel -m 0 and jetson_clocks.