I got a saved model converted to onnx in order to run inference using TensorRT c++ api;
but ouput results are different compared to python inference and I don’t why. seems like values are slightly shifted.
- tensorrt 8.2
- cuDNN 8.2
- onnx 1.8
- cuda 11,
- and did not set the BuilderFlag::kFP16 because my machine does not have that.
any help will be more than appreciated!
Nvidia Driver Version:
Operating System + Version:
Python Version (if applicable):
TensorFlow Version (if applicable):
PyTorch Version (if applicable):
Baremetal or Container (if container which image + tag):
Please attach or include links to any models, data, files, or scripts necessary to reproduce your issue. (Github repo, Google Drive, Dropbox, etc.)
- Exact steps/commands to build your repro
- Exact steps/commands to run your repro
- Full traceback of errors encountered