Hi,
I used both of trtexec and tensorrt Python API (build_cuda_engine()) to build TensorRT engine from ONNX models, but once a generate the TRT engine i can’t see the optimisations and how TensorRT edited the neuronal network before running the inference, i’am wondering if there is a tool which allow to see the optimisations done by TensorRT, if notis there a tool for profiling or to see how TensorRT do the INT8 quantization, deleting output and deleting layers etc…
Thanks