&&&& RUNNING TensorRT.trtexec [TensorRT v8400] # /usr/src/tensorrt/bin/trtexec --onnx=/home/tss-orin/Desktop/Arasu_test/TensorRT/quickstart/IntroNotebooks/Bhargav/Resnet101/resnet101_pytorch_BS_1.onnx [11/18/2022-11:43:20] [I] === Model Options === [11/18/2022-11:43:20] [I] Format: ONNX [11/18/2022-11:43:20] [I] Model: /home/tss-orin/Desktop/Arasu_test/TensorRT/quickstart/IntroNotebooks/Bhargav/Resnet101/resnet101_pytorch_BS_1.onnx [11/18/2022-11:43:20] [I] Output: [11/18/2022-11:43:20] [I] === Build Options === [11/18/2022-11:43:20] [I] Max batch: explicit batch [11/18/2022-11:43:20] [I] Memory Pools: workspace: default, dlaSRAM: default, dlaLocalDRAM: default, dlaGlobalDRAM: default [11/18/2022-11:43:20] [I] minTiming: 1 [11/18/2022-11:43:20] [I] avgTiming: 8 [11/18/2022-11:43:20] [I] Precision: FP32 [11/18/2022-11:43:20] [I] LayerPrecisions: [11/18/2022-11:43:20] [I] Calibration: [11/18/2022-11:43:20] [I] Refit: Disabled [11/18/2022-11:43:20] [I] Sparsity: Disabled [11/18/2022-11:43:20] [I] Safe mode: Disabled [11/18/2022-11:43:20] [I] DirectIO mode: Disabled [11/18/2022-11:43:20] [I] Restricted mode: Disabled [11/18/2022-11:43:20] [I] Build only: Disabled [11/18/2022-11:43:20] [I] Save engine: [11/18/2022-11:43:20] [I] Load engine: [11/18/2022-11:43:20] [I] Profiling verbosity: 0 [11/18/2022-11:43:20] [I] Tactic sources: Using default tactic sources [11/18/2022-11:43:20] [I] timingCacheMode: local [11/18/2022-11:43:20] [I] timingCacheFile: [11/18/2022-11:43:20] [I] Input(s)s format: fp32:CHW [11/18/2022-11:43:20] [I] Output(s)s format: fp32:CHW [11/18/2022-11:43:20] [I] Input build shapes: model [11/18/2022-11:43:20] [I] Input calibration shapes: model [11/18/2022-11:43:20] [I] === System Options === [11/18/2022-11:43:20] [I] Device: 0 [11/18/2022-11:43:20] [I] DLACore: [11/18/2022-11:43:20] [I] Plugins: [11/18/2022-11:43:20] [I] === Inference Options === [11/18/2022-11:43:20] [I] Batch: Explicit [11/18/2022-11:43:20] [I] Input inference shapes: model [11/18/2022-11:43:20] [I] Iterations: 10 [11/18/2022-11:43:20] [I] Duration: 3s (+ 200ms warm up) [11/18/2022-11:43:20] [I] Sleep time: 0ms [11/18/2022-11:43:20] [I] Idle time: 0ms [11/18/2022-11:43:20] [I] Streams: 1 [11/18/2022-11:43:20] [I] ExposeDMA: Disabled [11/18/2022-11:43:20] [I] Data transfers: Enabled [11/18/2022-11:43:20] [I] Spin-wait: Disabled [11/18/2022-11:43:20] [I] Multithreading: Disabled [11/18/2022-11:43:20] [I] CUDA Graph: Disabled [11/18/2022-11:43:20] [I] Separate profiling: Disabled [11/18/2022-11:43:20] [I] Time Deserialize: Disabled [11/18/2022-11:43:20] [I] Time Refit: Disabled [11/18/2022-11:43:20] [I] Inputs: [11/18/2022-11:43:20] [I] === Reporting Options === [11/18/2022-11:43:20] [I] Verbose: Disabled [11/18/2022-11:43:20] [I] Averages: 10 inferences [11/18/2022-11:43:20] [I] Percentile: 99 [11/18/2022-11:43:20] [I] Dump refittable layers:Disabled [11/18/2022-11:43:20] [I] Dump output: Disabled [11/18/2022-11:43:20] [I] Profile: Disabled [11/18/2022-11:43:20] [I] Export timing to JSON file: [11/18/2022-11:43:20] [I] Export output to JSON file: [11/18/2022-11:43:20] [I] Export profile to JSON file: [11/18/2022-11:43:20] [I] ==PROF== Connected to process 4094 (/usr/src/tensorrt/bin/trtexec) [11/18/2022-11:43:20] [I] === Device Information === [11/18/2022-11:43:20] [I] Selected Device: Orin [11/18/2022-11:43:20] [I] Compute Capability: 8.7 [11/18/2022-11:43:20] [I] SMs: 16 [11/18/2022-11:43:20] [I] Compute Clock Rate: 1.3 GHz [11/18/2022-11:43:20] [I] Device Global Memory: 30622 MiB [11/18/2022-11:43:20] [I] Shared Memory per SM: 164 KiB [11/18/2022-11:43:20] [I] Memory Bus Width: 128 bits (ECC disabled) [11/18/2022-11:43:20] [I] Memory Clock Rate: 1.3 GHz [11/18/2022-11:43:20] [I] [11/18/2022-11:43:20] [I] TensorRT version: 8.4.0 [11/18/2022-11:43:21] [I] [TRT] [MemUsageChange] Init CUDA: CPU +522, GPU +0, now: CPU 548, GPU 5832 (MiB) [11/18/2022-11:43:24] [I] [TRT] [MemUsageChange] Init builder kernel library: CPU +584, GPU +549, now: CPU 1152, GPU 6399 (MiB) [11/18/2022-11:43:24] [I] Start parsing network model [11/18/2022-11:43:24] [I] [TRT] ---------------------------------------------------------------- [11/18/2022-11:43:24] [I] [TRT] Input filename: /home/tss-orin/Desktop/Arasu_test/TensorRT/quickstart/IntroNotebooks/Bhargav/Resnet101/resnet101_pytorch_BS_1.onnx [11/18/2022-11:43:24] [I] [TRT] ONNX IR version: 0.0.4 [11/18/2022-11:43:24] [I] [TRT] Opset version: 9 [11/18/2022-11:43:24] [I] [TRT] Producer name: pytorch [11/18/2022-11:43:24] [I] [TRT] Producer version: 1.12.0 [11/18/2022-11:43:24] [I] [TRT] Domain: [11/18/2022-11:43:24] [I] [TRT] Model version: 0 [11/18/2022-11:43:24] [I] [TRT] Doc string: [11/18/2022-11:43:24] [I] [TRT] ---------------------------------------------------------------- [11/18/2022-11:43:24] [I] Finish parsing network model [11/18/2022-11:43:24] [I] [TRT] ---------- Layers Running on DLA ---------- [11/18/2022-11:43:24] [I] [TRT] ---------- Layers Running on GPU ---------- [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_0 + Relu_1 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] POOLING: MaxPool_2 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_3 + Relu_4 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_5 + Relu_6 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_8 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_7 + Add_9 + Relu_10 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_11 + Relu_12 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_13 + Relu_14 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_15 + Add_16 + Relu_17 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_18 + Relu_19 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_20 + Relu_21 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_22 + Add_23 + Relu_24 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_25 + Relu_26 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_27 + Relu_28 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_30 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_29 + Add_31 + Relu_32 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_33 + Relu_34 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_35 + Relu_36 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_37 + Add_38 + Relu_39 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_40 + Relu_41 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_42 + Relu_43 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_44 + Add_45 + Relu_46 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_47 + Relu_48 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_49 + Relu_50 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_51 + Add_52 + Relu_53 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_54 + Relu_55 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_56 + Relu_57 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_59 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_58 + Add_60 + Relu_61 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_62 + Relu_63 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_64 + Relu_65 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_66 + Add_67 + Relu_68 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_69 + Relu_70 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_71 + Relu_72 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_73 + Add_74 + Relu_75 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_76 + Relu_77 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_78 + Relu_79 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_80 + Add_81 + Relu_82 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_83 + Relu_84 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_85 + Relu_86 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_87 + Add_88 + Relu_89 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_90 + Relu_91 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_92 + Relu_93 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_94 + Add_95 + Relu_96 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_97 + Relu_98 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_99 + Relu_100 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_101 + Add_102 + Relu_103 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_104 + Relu_105 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_106 + Relu_107 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_108 + Add_109 + Relu_110 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_111 + Relu_112 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_113 + Relu_114 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_115 + Add_116 + Relu_117 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_118 + Relu_119 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_120 + Relu_121 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_122 + Add_123 + Relu_124 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_125 + Relu_126 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_127 + Relu_128 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_129 + Add_130 + Relu_131 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_132 + Relu_133 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_134 + Relu_135 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_136 + Add_137 + Relu_138 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_139 + Relu_140 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_141 + Relu_142 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_143 + Add_144 + Relu_145 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_146 + Relu_147 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_148 + Relu_149 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_150 + Add_151 + Relu_152 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_153 + Relu_154 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_155 + Relu_156 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_157 + Add_158 + Relu_159 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_160 + Relu_161 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_162 + Relu_163 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_164 + Add_165 + Relu_166 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_167 + Relu_168 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_169 + Relu_170 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_171 + Add_172 + Relu_173 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_174 + Relu_175 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_176 + Relu_177 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_178 + Add_179 + Relu_180 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_181 + Relu_182 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_183 + Relu_184 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_185 + Add_186 + Relu_187 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_188 + Relu_189 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_190 + Relu_191 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_192 + Add_193 + Relu_194 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_195 + Relu_196 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_197 + Relu_198 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_199 + Add_200 + Relu_201 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_202 + Relu_203 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_204 + Relu_205 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_206 + Add_207 + Relu_208 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_209 + Relu_210 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_211 + Relu_212 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_213 + Add_214 + Relu_215 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_216 + Relu_217 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_218 + Relu_219 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_221 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_220 + Add_222 + Relu_223 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_224 + Relu_225 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_226 + Relu_227 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_228 + Add_229 + Relu_230 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_231 + Relu_232 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_233 + Relu_234 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Conv_235 + Add_236 + Relu_237 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] POOLING: GlobalAveragePool_238 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] CONVOLUTION: Gemm_240 [11/18/2022-11:43:24] [I] [TRT] [GpuLayer] SHUFFLE: reshape_after_Gemm_240 [11/18/2022-11:43:25] [I] [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +962, GPU +747, now: CPU 2285, GPU 7468 (MiB) [11/18/2022-11:43:26] [I] [TRT] [MemUsageChange] Init cuDNN: CPU +134, GPU +201, now: CPU 2419, GPU 7669 (MiB) [11/18/2022-11:43:26] [I] [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored. ==ERROR== Failed to prepare kernel for profiling ==ERROR== Failed to profile kernel "copyPackedKernel" in process 4094 ==PROF== Trying to shutdown target application ==ERROR== The application returned an error code (9). ==ERROR== An error occurred while trying to profile. ==WARNING== No kernels were profiled. ==WARNING== Profiling kernels launched by child processes requires the --target-processes all option.