I am transferring my onnx
model to trt
model on agx.
I successfully dis that on jp4.4 trt 7 with the following command.
./trtexec --onnx=xxx1.onnx --saveEngine=xxx.trt --minShapes=input_1:0:1x224x224x1 --optShapes=input_1:0:2x224x224x1 --maxShapes=input_1:0:2x224x224x1 --workspace=4096 --verbose --fp16
But when i turn to trt 8.2, jp4.6 with the same command it throws out message like below
[09/07/2022-09:09:28] [I] Engine built in 240.911 sec.
[09/07/2022-09:09:28] [V] [TRT] Using cublas as a tactic source
[09/07/2022-09:09:28] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +0, GPU +0, now: CPU 1384, GPU 17322 (MiB)
[09/07/2022-09:09:28] [V] [TRT] Using cuDNN as a tactic source
[09/07/2022-09:09:28] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +0, GPU +0, now: CPU 1384, GPU 17322 (MiB)
[09/07/2022-09:09:28] [V] [TRT] Total per-runner device persistent memory is 421888
[09/07/2022-09:09:28] [V] [TRT] Total per-runner host persistent memory is 39168
[09/07/2022-09:09:28] [V] [TRT] Allocated activation device memory of size 1608192
[09/07/2022-09:09:28] [I] [TRT] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +2, now: CPU 0, GPU 2 (MiB)
[09/07/2022-09:09:28] [I] Using random values for input input_1
[09/07/2022-09:09:28] [I] Created input binding for input_1 with dimensions 2x224x224x1
[09/07/2022-09:09:28] [I] Using random values for output concat
[09/07/2022-09:09:28] [I] Created output binding for concat with dimensions 2x12
[09/07/2022-09:09:28] [I] Using random values for output concat_before_shuffle
[09/07/2022-09:09:28] [I] Created output binding for concat_before_shuffle with dimensions -1x12x1x1
[09/07/2022-09:09:28] [I] Starting inference
[09/07/2022-09:09:28] [E] Error[1]: [genericReformat.cu::executeMemcpy::1334] Error Code 1: Cuda Runtime (invalid argument)
[09/07/2022-09:09:28] [E] Error occurred during inference
&&&& FAILED TensorRT.trtexec [TensorRT v8201] #
Here is my onnx
model
test_nvidia.onnx (708.1 KB)