Calls tensorrt inference at 10 fps, the cost time per call is 30ms.
But calls at 50 fps, it is down to 13ms, reduced so much.
How is the inference cost time affected by the frequency of calls?
TensorRT Version: TensorRT-22.214.171.124
GPU Type: Tesla 4
Nvidia Driver Version: 418.87.00
CUDA Version: 10.2
Operating System + Version: Ubuntu18.04
Python Version (if applicable):
TensorFlow Version (if applicable):
PyTorch Version (if applicable): 1.4
Baremetal or Container (if container which image + tag):