TensorRT

Description

I created a simple model on mnist dataset and made a TRT engine as given in the TensorRT/samples/python/network_api_pytorch_mnist . The model is performing well ,while i use the model.predict method in Tensorflow,But after converting the model Into TRT engine,The performance is poor.

I will attach mine and Nvidia GIT repository (The exam which i tried to replicate) ,Can anyone please take a lot and help me out.

Environment

TensorRT Version: 8.0
GPU Type: GTX 1050 TI
Nvidia Driver Version: 470.57.02
CUDA Version: 11.4
CUDNN Version: 8.2
Operating System + Version: Ubuntu 20.04
Python Version (if applicable): 3.8
TensorFlow Version (if applicable): 2.5
PyTorch Version (if applicable):
Baremetal or Container (if container which image + tag):

Relevant Files

My Repository

Nvidia-TensorRT example Repository

Steps To Reproduce

Please include:

  • Exact steps/commands to build your repro
  • Exact steps/commands to run your repro
  • Full traceback of errors encountered

Hi @Ragu_23
Please refer to below thread in case it’s helpful in your case:
https://docs.nvidia.com/deeplearning/tensorrt/best-practices/index.html#optimize-python

Also, could you please share the performance results of both TRT and TF results in this case?

Thanks