ek9852
June 28, 2020, 11:42pm
1
Description
createInferRuntime api call blocked for ~ 8 seconds with 100% cpu loading on single cpu.
Environment
TensorRT Version : nvcr.io/nvidia/tensorrt 20.06-py3
GPU Type : Tesla T4
Nvidia Driver Version : 418.87.00
CUDA Version : 11.0
CUDNN Version :
Operating System + Version : Linux docker on nvcr.io/nvidia/tensorrt 20.06-py3 . (Host amazon ecs-gpu ami)
Steps To Reproduce
Write a C program and just call createInferRuntime(gLogger); iusing C API.
It takes ~8seconds for the call to finish.
Expected to be much faster.
Hi @ek9852 ,
You can try reproducing the issue after upgrading Nvidia Driver Version, as the latest one is r450.
Please find the support matrix to check for the same.
These support matrices provide a look into the supported versions of the OS, NVIDIA CUDA, the CUDA driver, and the hardware for the NVIDIA cuDNN 8.5.0 release.
Support Matrix :: NVIDIA Deep Learning TensorRT Documentation .
Please share your code in case if the issue persist.
Thanks!
ek9852
July 24, 2020, 10:48pm
3
Tried again on Tesla T5
NVIDIA-SMI 450.51.05 Driver Version: 450.51.05 CUDA Version: 11.0
Same problem.createInferRuntime takes more than 10 seconds on Tesla T4.
While in TITAN X createInferRuntime takes ~ 1-2 seconds.
1 Like