My device is Jetson AGX, jetpack 4.6,teensorrt8.2. My question is why tensorrt runs with very high latency after I upgraded my kernel from 32.7.2 to 32.7.4.
before I upgrade the kernel(32.7.2),
I run the tensorrt inference CPP demo app. I got the below profiler.
after I upgraded the kernel from 32.7.2 to 32.7.4, the same app same input but the latency was very high.
the tensorrt API “context->setBindingDimensions(0, inferInputDims);” is very slow.
why ??? the context->setBindingDimensions(0, inferInputDims);running so slow ?What kind of resources does it take up? What resources does it require to operate?
I run the same demo and inputs on the same tensorrt version.
but the kernel version is not the same; I am curious whether the kernel imports some bugs that affect the tensorrt inference.
why ??? the context->setBindingDimensions(0, inferInputDims); running so slow ?What kind of resources does it take up? What resources does it require to operate?
The function is just set up a parameter that should be fast.
Based on the results you provided, we don’t see obvious latency degradation of the SetBindingDimensions.