Description
Hi there
i got a saved model converted to onnx in order to run inference using Tensorrt 8.6 c++ API.
The model is a virtual model from the openai clips model.
(https://clip-as-service.s3.us-east-2.amazonaws.com/models-436c69702d61732d53657276696365/onnx/ViT-B-32/visual.onnx)
checking the embedded value of the image in the model.
but output results are diffrence compared to python
infrence(pt or onnx runtime) to trt in tensorrt c++ api
I checked and found that python inference is correct and tensorrt c++ api is incorrect. (both fp32)
The next log is trtexec verbose log after onnx check. You can check it (
https://drive.google.com/file/d/1carAjQ_oP2xEkia48J0tNztQ5meukaqZ/view?usp=drive_link)
Environment
TensorRT Version: 8.6.1GA
GPU Type: GTX3090
Nvidia Driver Version:
CUDA Version: 12.0
CUDNN Version: 8.9.4.25
Operating System + Version:
Python Version (if applicable): 3.10.1
TensorFlow Version (if applicable):
PyTorch Version (if applicable): 2.0
Baremetal or Container (if container which image + tag):
Relevant Files
Please attach or include links to any models, data, files, or scripts necessary to reproduce your issue. (Github repo, Google Drive, Dropbox, etc.)
Steps To Reproduce
Please include:
- Exact steps/commands to build your repro
- Exact steps/commands to run your repro
- Full traceback of errors encountered