Hi,
Request you to share the ONNX model and the script if not shared already so that we can assist you better.
Alongside you can try few things:
validating your model with the below snippet
check_model.py
import sys
import onnx
filename = yourONNXmodel
model = onnx.load(filename)
onnx.checker.check_model(model).
2) Try running your model with trtexec command.
In case you are still facing issue, request you to share the trtexec “”–verbose"" log for further debugging
Thanks!
Serialization of the model saves time when you plan to reuse the model with TensorRT on the same platform.
You need not rebuild every time, which saves time. Yes, for the transmission you need to use the ONNX model.
TensorRT built engine is specific to the platform(GPU, OS, TensorRT version, etc) and not portable across the different platforms.