[TensorRT] INTERNAL ERROR: Assertion failed: cublasStatus == CUBLAS_STATUS_SUCCESS ../rtSafe/cublas/cublasLtWrapper.cpp:279


CenterNet ONNX to Tensorrt, low version is no problem but meet a bug in high version.

[TensorRT] INTERNAL ERROR: Assertion failed: cublasStatus == CUBLAS_STATUS_SUCCESS
[TensorRT] ERROR: …/rtSafe/cublas/cublasLtWrapper.cpp (279) - Assertion Error in getCublasLtHeuristic: 0 (cublasStatus == CUBLAS_STATUS_SUCCESS)
ERROR: failed to build the TensorRT engine!


TensorRT Version:
GPU Type: 1660Ti
Nvidia Driver Version: 455.23.05
CUDA Version: 11.1
CUDNN Version: 8.04
Operating System + Version: ubantu 18.04
Python Version (if applicable): 3.6
TensorFlow Version (if applicable):
PyTorch Version (if applicable): 1.7.1
Baremetal or Container (if container which image + tag):

Relevant Files

Hi @lcuknthing,

We are able to successfully convert ONNX model to trt, using trtexec command as following.
trtexec --onnx=20210324133230_fix_size.onnx --verbose --explicitBatch --shapes=input_image:1x3x448x448
I tried on trt version :

We request you to validate TensorRT installation steps you followed, for your reference please check installation guide.

We recommend you to try on TensorRT NGC container to avoid system dependencies.

Thank you.