Hi,
Thank you so much for your reply!
Long story short, something messed up and I ended up with this error.
I am now using another GPU (CUDA 12.5) and can convert ONNX to TensorRT
Is there a code or documentation I can refer on how to use the “.trt” file to run inference?
Thanks again :))