Hi,
Request you to share the ONNX model and the script if not shared already so that we can assist you better.
Alongside you can try few things:
validating your model with the below snippet
check_model.py
import sys
import onnx
filename = yourONNXmodel
model = onnx.load(filename)
onnx.checker.check_model(model).
2) Try running your model with trtexec command. https://github.com/NVIDIA/TensorRT/tree/master/samples/opensource/trtexec
In case you are still facing issue, request you to share the trtexec “”–verbose"" log for further debugging
Thanks!
Based on error, looks like you’re using Double data type in your model. Old version of TensorRT doesn’t support Double data type. On latest TRT version 8.0, we have support for Double data type. On latest version we are unable to reproduce this issue and able to successfully build TRT engine.
Please refer onnx-tensorrt/operators.md at main · onnx/onnx-tensorrt · GitHub for more details.