Description
Calls tensorrt inference at 10 fps, the cost time per call is 30ms.
But calls at 50 fps, it is down to 13ms, reduced so much.
How is the inference cost time affected by the frequency of calls?
Environment
TensorRT Version: TensorRT-7.1.3.4
GPU Type: Tesla 4
Nvidia Driver Version: 418.87.00
CUDA Version: 10.2
CUDNN Version:
Operating System + Version: Ubuntu18.04
Python Version (if applicable):
TensorFlow Version (if applicable):
PyTorch Version (if applicable): 1.4
Baremetal or Container (if container which image + tag):