Description
For an image generator (SDXL)
I am experiencing issues while building and running a TensorRT engine on a NVIDIA RTX 5090 GPU with CUDA 12.8.
The conversion of models (e.g., PyTorch to ONNX) and subsequent inference execution fails due to compatibility issues between the latest CUDA version and TensorRT.
This causes errors such as model export failures, dynamic shape issues, and performance degradation.
Etc.
GPT-Research / GPT-Pro failed to help, suggesting hacks and fixed I consider as ‘daring’ and ‘wild’, ending up generating ‘broken’ images with tensorrt.
The official Comfy-UI tensorrt nodes are not working either.
Environment
TensorRT Version: cu12_bindings-10.14.1.48.post1-cp312-none-manylinux_2_28_x86_64.whl
GPU Type: NVIDIA RTX 5090
Nvidia Driver Version: 570.195.03
CUDA Version: 12.8
CUDNN Version: cudnn-cu12==9.10.2.21
Operating System + Version: Ubuntu / Linux on Runpod
Python Version: python3.12
PyTorch Version: 2.10.0.dev20250924+cu128
Baremetal or Container (if container which image + tag): (hearmeman/comfyui-wan-template:v10) on Runpod
Relevant Files
Please attach or include links to any models, data, files, or scripts necessary to reproduce your issue. (Github repo, Google Drive, Dropbox, etc.)
Steps To Reproduce
Step 1:
Buy RTX5090 / Install Cuda 12.8, + setup Comfy ui (or use runpod hearmeman/comfyui-wan-template:v10)
Step 2:
Install and run official Tensorrt from Comfy-UI Manager.
Step 3:
You will see Error.