Trtexec log problem and use DLA error on Jetson Xavier

Hi,

I have run my onnx model successfully using

./trtexec --onnx=/home/nvidia/APA_FS_Model.onnx --explicitBatch --verbose

[02/19/2021-11:22:18] [I] Starting inference threads
[02/19/2021-11:22:21] [I] Warmup completed 0 queries over 200 ms
[02/19/2021-11:22:21] [I] Timing trace has 0 queries over 3.00517 s
[02/19/2021-11:22:21] [I] Trace averages of 10 runs:

[02/19/2021-11:22:21] [I] Host Latency
[02/19/2021-11:22:21] [I] min: 1.92184 ms (end to end 1.92749 ms)
[02/19/2021-11:22:21] [I] max: 1.99609 ms (end to end 2.00391 ms)
[02/19/2021-11:22:21] [I] mean: 1.96004 ms (end to end 1.96797 ms)
[02/19/2021-11:22:21] [I] median: 1.9613 ms (end to end 1.96899 ms)
[02/19/2021-11:22:21] [I] percentile: 1.98669 ms at 99% (end to end 1.99561 ms at 99%)
[02/19/2021-11:22:21] [I] throughput: 0 qps
[02/19/2021-11:22:21] [I] walltime: 3.00517 s
[02/19/2021-11:22:21] [I] Enqueue Time
[02/19/2021-11:22:21] [I] min: 0.672119 ms
[02/19/2021-11:22:21] [I] max: 1.63806 ms
[02/19/2021-11:22:21] [I] median: 0.804443 ms
[02/19/2021-11:22:21] [I] GPU Compute
[02/19/2021-11:22:21] [I] min: 1.83517 ms
[02/19/2021-11:22:21] [I] max: 1.90869 ms
[02/19/2021-11:22:21] [I] mean: 1.87262 ms
[02/19/2021-11:22:21] [I] median: 1.87378 ms
[02/19/2021-11:22:21] [I] percentile: 1.89844 ms at 99%
[02/19/2021-11:22:21] [I] total compute time: 2.85949 s

Then I tried to use DLA to run trtexec using my onnx model with command below:

./trtexec --onnx=/home/nvidia/APA_FS_Model.onnx --explicitBatch --verbose --useDLACore=0 --allowGPUFallback --workspace=2048

It gives message :

[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer (Unnamed Layer* 0) [Constant] is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer (Unnamed Layer* 1) [Shuffle] is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] Conv_26: DLA only supports prePadding < kernelSize. Detected: [4,4] < [3,3].
[02/19/2021-11:10:13] [W] [TRT] Conv_26: DLA only supports postPadding < kernelSize. Detected: [4,4] < [3,3].
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer Conv_26 is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] Conv_34: DLA only supports prePadding < kernelSize. Detected: [4,4] < [3,3].
[02/19/2021-11:10:13] [W] [TRT] Conv_34: DLA only supports postPadding < kernelSize. Detected: [4,4] < [3,3].
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer Conv_34 is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] ConvTranspose_49: DLA cores do not support more than 1 groups.
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer ConvTranspose_49 is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] ConvTranspose_57: DLA cores do not support more than 1 groups.
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer ConvTranspose_57 is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [W] [TRT] ConvTranspose_60: DLA cores do not support more than 1 groups.
[02/19/2021-11:10:13] [W] [TRT] Default DLA is enabled but layer ConvTranspose_60 is not supported on DLA, falling back to GPU.
[02/19/2021-11:10:13] [V] [TRT] Applying generic optimizations to the graph for inference.
[02/19/2021-11:10:13] [V] [TRT] Original: 62 layers
[02/19/2021-11:10:13] [V] [TRT] After dead-layer removal: 62 layers
[02/19/2021-11:10:13] [W] [TRT] DLA allows only same dimensions inputs to Elementwise.
[02/19/2021-11:10:13] [W] [TRT] Internal DLA error for layer Add_1. Switching to GPU fallback.
[02/19/2021-11:10:13] [V] [TRT] After DLA optimization: 13 layers
[02/19/2021-11:10:13] [V] [TRT] Fusing (Unnamed Layer* 0) [Constant] with (Unnamed Layer* 1) [Shuffle]
[02/19/2021-11:10:13] [V] [TRT] After Myelin optimization: 12 layers
[02/19/2021-11:10:13] [V] [TRT] Fusing (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] with Add_1
[02/19/2021-11:10:13] [V] [TRT] After scale fusion: 11 layers
[02/19/2021-11:10:13] [V] [TRT] After vertical fusions: 11 layers
[02/19/2021-11:10:13] [V] [TRT] After final dead-layer removal: 11 layers
[02/19/2021-11:10:13] [V] [TRT] After tensor merging: 11 layers
[02/19/2021-11:10:13] [V] [TRT] After concat removal: 11 layers
[02/19/2021-11:10:13] [V] [TRT] Graph construction and optimization completed in 0.0485818 seconds.
[02/19/2021-11:10:13] [I] [TRT]
[02/19/2021-11:10:13] [I] [TRT] --------------- Layers running on DLA:
[02/19/2021-11:10:13] [I] [TRT] {Conv_2,Relu_3,Conv_4,Relu_5,Conv_6,Conv_7,Relu_8,Add_9,Conv_10,Relu_11,Conv_12,Conv_13,Relu_14,Add_15,Conv_16,Relu_17,Conv_18,Relu_19,Conv_20,Relu_21,Add_22,Conv_23,Relu_24,Add_25,Add_41,Add_42,Conv_50,Relu_51}, {Relu_27,Conv_28,Relu_29,Add_30,Conv_31,Relu_32,Add_33,Add_43,Add_44}, {Relu_35,Conv_36,Relu_37,Add_38,Conv_39,Relu_40,Add_45,Add_46,Conv_47,Relu_48}, {Add_52,Conv_53,Relu_54,Conv_55,Relu_56}, {Conv_58,Relu_59},
[02/19/2021-11:10:13] [I] [TRT] --------------- Layers running on GPU:
[02/19/2021-11:10:13] [I] [TRT] (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] + Add_1, Conv_26, Conv_34, ConvTranspose_49, ConvTranspose_57, ConvTranspose_60,
[02/19/2021-11:10:15] [V] [TRT] Constructing optimization profile number 0 [1/1].
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 1002 time 0.084996
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.124564
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 1002 Time: 0.084996
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 1002 time 0.90214
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.064912
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.064912
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 1002 time 1.11668
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.083608
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.083608
[02/19/2021-11:10:15] [V] [TRT] *************** Autotuning format combination: Float(1,256,49152,147456) → Float(1,256,49152,147456) ***************
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] + Add_1 (Scale)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.058492
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.058492
[02/19/2021-11:10:15] [V] [TRT] *************** Autotuning format combination: Half(1,256,49152,147456) → Half(1,256,49152,147456) ***************
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] + Add_1 (Scale)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.04428
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.04428
[02/19/2021-11:10:15] [V] [TRT] *************** Autotuning format combination: Half(1,256,49152:2,98304) → Half(1,256,49152:2,98304) ***************
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] + Add_1 (Scale)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.049664
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.049664
[02/19/2021-11:10:15] [V] [TRT] *************** Autotuning format combination: Half(1,256,1:8,49152) → Half(1,256,1:8,49152) ***************
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Unnamed Layer* 0) [Constant] + (Unnamed Layer* 1) [Shuffle] + Add_1 (Scale)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.087672
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 0 Time: 0.087672
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:15] [V] [TRT] Tactic: 1002 time 0.0699
[02/19/2021-11:10:15] [V] [TRT] Tactic: 0 time 0.12094
[02/19/2021-11:10:15] [V] [TRT] Fastest Tactic: 1002 Time: 0.0699
[02/19/2021-11:10:15] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 0.913276
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.060428
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.060428
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 1.10758
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.08
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.08
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 1.10501
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.079872
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.079872
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 0.90982
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.951668
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 1002 Time: 0.90982
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 0.072056
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.102384
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 1002 Time: 0.072056
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 0.911372
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.058892
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.058892
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 1.10722
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.079364
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.079364
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 1.11476
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.080028
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 0 Time: 0.080028
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:16] [V] [TRT] Tactic: 1002 time 0.91096
[02/19/2021-11:10:16] [V] [TRT] Tactic: 0 time 0.941792
[02/19/2021-11:10:16] [V] [TRT] Fastest Tactic: 1002 Time: 0.91096
[02/19/2021-11:10:16] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.195576
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.050164
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.050164
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.197008
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.041996
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.041996
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 1.10823
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.102376
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.102376
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 1.11014
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.057864
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.057864
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.910964
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.944124
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 1002 Time: 0.910964
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.213516
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.060156
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.060156
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.212868
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.050828
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.050828
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 1.10364
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.053248
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.053248
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.90702
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 0.101624
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 0 Time: 0.101624
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:17] [V] [TRT] Tactic: 1002 time 0.905732
[02/19/2021-11:10:17] [V] [TRT] Tactic: 0 time 1.00557
[02/19/2021-11:10:17] [V] [TRT] Fastest Tactic: 1002 Time: 0.905732
[02/19/2021-11:10:17] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 1.11129
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.08026
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 0 Time: 0.08026
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.916348
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.953356
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 1002 Time: 0.916348
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 1.11206
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.079496
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 0 Time: 0.079496
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.9248
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.940924
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 1002 Time: 0.9248
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 1.11117
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.103452
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 0 Time: 0.103452
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.916328
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.944916
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 1002 Time: 0.916328
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.904968
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.966268
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 1002 Time: 0.904968
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.352272
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.040588
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 0 Time: 0.040588
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.351812
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.391624
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 1002 Time: 0.351812
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: (Reformat)
[02/19/2021-11:10:18] [V] [TRT] Tactic: 1002 time 0.351968
[02/19/2021-11:10:18] [V] [TRT] Tactic: 0 time 0.041632
[02/19/2021-11:10:18] [V] [TRT] Fastest Tactic: 0 Time: 0.041632
[02/19/2021-11:10:18] [V] [TRT] *************** Autotuning format combination: Half(1,256,1:4,49152) → Half(1,32,768:16,3072), Half(1,32,768:16,3072), Half(1,64,3072:16,9216) ***************
[02/19/2021-11:10:18] [V] [TRT] --------------- Timing Runner: {Conv_2,Relu_3,Conv_4,Relu_5,Conv_6,Conv_7,Relu_8,Add_9,Conv_10,Relu_11,Conv_12,Conv_13,Relu_14,Add_15,Conv_16,Relu_17,Conv_18,Relu_19,Conv_20,Relu_21,Add_22,Conv_23,Relu_24,Add_25,Add_41,Add_42,Conv_50,Relu_51} (DLA)
[02/19/2021-11:10:19] [W] [TRT] DLA Node compilation Failed.
[02/19/2021-11:10:19] [V] [TRT] Tactic: 549596028979 skipped. ProcessNode failure.
[02/19/2021-11:10:19] [V] [TRT] Fastest Tactic: -3360065831133338131 Time: 3.40282e+38
[02/19/2021-11:10:19] [V] [TRT] *************** Autotuning format combination: Half(1,256,49152:16,49152) → Half(1,32,768:16,3072), Half(1,32,768:16,3072), Half(1,64,3072:16,9216) ***************
[02/19/2021-11:10:19] [V] [TRT] --------------- Timing Runner: {Conv_2,Relu_3,Conv_4,Relu_5,Conv_6,Conv_7,Relu_8,Add_9,Conv_10,Relu_11,Conv_12,Conv_13,Relu_14,Add_15,Conv_16,Relu_17,Conv_18,Relu_19,Conv_20,Relu_21,Add_22,Conv_23,Relu_24,Add_25,Add_41,Add_42,Conv_50,Relu_51} (DLA)
[02/19/2021-11:10:19] [W] [TRT] DLA Node compilation Failed.
[02/19/2021-11:10:19] [V] [TRT] Tactic: 549596028979 skipped. ProcessNode failure.
[02/19/2021-11:10:19] [V] [TRT] Fastest Tactic: -3360065831133338131 Time: 3.40282e+38
[02/19/2021-11:10:19] [E] [TRT] Try increasing the workspace size with IBuilderConfig::setMaxWorkspaceSize() if using IBuilder::buildEngineWithConfig, or IBuilder::setMaxWorkspaceSize() if using IBuilder::buildCudaEngine.
[02/19/2021-11:10:19] [E] [TRT] …/builder/tacticOptimizer.cpp (1715) - TRTInternal Error in computeCosts: 0 (Could not find any implementation for node {Conv_2,Relu_3,Conv_4,Relu_5,Conv_6,Conv_7,Relu_8,Add_9,Conv_10,Relu_11,Conv_12,Conv_13,Relu_14,Add_15,Conv_16,Relu_17,Conv_18,Relu_19,Conv_20,Relu_21,Add_22,Conv_23,Relu_24,Add_25,Add_41,Add_42,Conv_50,Relu_51}.)
[02/19/2021-11:10:19] [V] [TRT] Builder timing cache: created 33 entries, 0 hit(s)
[02/19/2021-11:10:19] [E] [TRT] …/builder/tacticOptimizer.cpp (1715) - TRTInternal Error in computeCosts: 0 (Could not find any implementation for node {Conv_2,Relu_3,Conv_4,Relu_5,Conv_6,Conv_7,Relu_8,Add_9,Conv_10,Relu_11,Conv_12,Conv_13,Relu_14,Add_15,Conv_16,Relu_17,Conv_18,Relu_19,Conv_20,Relu_21,Add_22,Conv_23,Relu_24,Add_25,Add_41,Add_42,Conv_50,Relu_51}.)
[02/19/2021-11:10:19] [E] Engine creation failed
[02/19/2021-11:10:19] [E] Engine set up failed

How do I solve this problem?

Thanks

Environment

TensorRT Version : 7.1.3
GPU Type : Xavier
Nvidia Driver Version : Package:nvidia-jetpack, Version: 4.4.1-b50
CUDA Version : 10.2.89
CUDNN Version : 8.0.0
Operating System + Version : Ubuntu 18.04
Python Version (if applicable) :
TensorFlow Version (if applicable) :
PyTorch Version (if applicable) :
Baremetal or Container (if container which image + tag) :

Hi,

Would you mind to increase the workspace size to see if it helps?

./trtexec --onnx=/home/nvidia/APA_FS_Model.onnx --explicitBatch --verbose --useDLACore=0 --allowGPUFallback --workspace=4096

Thanks.

Thanks for reply.

I have tried

./trtexec --onnx=/home/nvidia/APA_FS_Model.onnx --explicitBatch --verbose --useDLACore=0 --allowGPUFallback --workspace=4096
and

./trtexec --onnx=/home/nvidia/APA_FS_Model.onnx --explicitBatch --verbose --useDLACore=0 --allowGPUFallback --workspace=8192

Unfortunately, It gives same error as usual.
Thanks.

Hi,

Could you share the ONNX model with us?
We want to reproduce this issue in our environment first.

Thanks.

Hi,

I have directly messaged the model to you.

Thanks for help.

Thanks.

We have downloaded your model.
Will let you know for any progress.

Hi,

Thanks for your patience.

It seems that you are facing a similar issue as Cannot build a TensorRT engine for DLA from a large ONNX file.
This issue is fixed internally and will be included in our next release.
Please wait for our update.

Thanks.