[video] created imageLoader from file:///home/xavier/jetson-inference/build/aarch64/bin/images/strawberry_0.jpg ------------------------------------------------ imageLoader video options: ------------------------------------------------ -- URI: file:///home/xavier/jetson-inference/build/aarch64/bin/images/strawberry_0.jpg - protocol: file - location: images/strawberry_0.jpg - extension: jpg -- deviceType: file -- ioType: input -- codec: unknown -- width: 0 -- height: 0 -- frameRate: 0.000000 -- bitRate: 0 -- numBuffers: 4 -- zeroCopy: true -- flipMethod: none -- loop: 0 -- rtspLatency 2000 ------------------------------------------------ [video] created imageWriter from file:///home/xavier/jetson-inference/build/aarch64/bin/images/test/output_1.jpg ------------------------------------------------ imageWriter video options: ------------------------------------------------ -- URI: file:///home/xavier/jetson-inference/build/aarch64/bin/images/test/output_1.jpg - protocol: file - location: images/test/output_1.jpg - extension: jpg -- deviceType: file -- ioType: output -- codec: unknown -- width: 0 -- height: 0 -- frameRate: 0.000000 -- bitRate: 0 -- numBuffers: 4 -- zeroCopy: true -- flipMethod: none -- loop: 0 -- rtspLatency 2000 ------------------------------------------------ [OpenGL] glDisplay -- X screen 0 resolution: 1920x1080 [OpenGL] glDisplay -- X window resolution: 1920x1080 [OpenGL] glDisplay -- display device initialized (1920x1080) [video] created glDisplay from display://0 ------------------------------------------------ glDisplay video options: ------------------------------------------------ -- URI: display://0 - protocol: display - location: 0 -- deviceType: display -- ioType: output -- codec: raw -- width: 1920 -- height: 1080 -- frameRate: 0.000000 -- bitRate: 0 -- numBuffers: 4 -- zeroCopy: true -- flipMethod: none -- loop: 0 -- rtspLatency 2000 ------------------------------------------------ imageNet -- loading classification network model from: -- prototxt networks/googlenet.prototxt -- model networks/bvlc_googlenet.caffemodel -- class_labels networks/ilsvrc12_synset_words.txt -- input_blob 'data' -- output_blob 'prob' -- batch_size 1 [TRT] TensorRT version 8.2.1 [TRT] loading NVIDIA plugins... [TRT] Registered plugin creator - ::GridAnchor_TRT version 1 [TRT] Registered plugin creator - ::GridAnchorRect_TRT version 1 [TRT] Registered plugin creator - ::NMS_TRT version 1 [TRT] Registered plugin creator - ::Reorg_TRT version 1 [TRT] Registered plugin creator - ::Region_TRT version 1 [TRT] Registered plugin creator - ::Clip_TRT version 1 [TRT] Registered plugin creator - ::LReLU_TRT version 1 [TRT] Registered plugin creator - ::PriorBox_TRT version 1 [TRT] Registered plugin creator - ::Normalize_TRT version 1 [TRT] Registered plugin creator - ::ScatterND version 1 [TRT] Registered plugin creator - ::RPROI_TRT version 1 [TRT] Registered plugin creator - ::BatchedNMS_TRT version 1 [TRT] Registered plugin creator - ::BatchedNMSDynamic_TRT version 1 [TRT] Could not register plugin creator - ::FlattenConcat_TRT version 1 [TRT] Registered plugin creator - ::CropAndResize version 1 [TRT] Registered plugin creator - ::DetectionLayer_TRT version 1 [TRT] Registered plugin creator - ::EfficientNMS_TRT version 1 [TRT] Registered plugin creator - ::EfficientNMS_ONNX_TRT version 1 [TRT] Registered plugin creator - ::EfficientNMS_TFTRT_TRT version 1 [TRT] Registered plugin creator - ::Proposal version 1 [TRT] Registered plugin creator - ::ProposalLayer_TRT version 1 [TRT] Registered plugin creator - ::PyramidROIAlign_TRT version 1 [TRT] Registered plugin creator - ::ResizeNearest_TRT version 1 [TRT] Registered plugin creator - ::Split version 1 [TRT] Registered plugin creator - ::SpecialSlice_TRT version 1 [TRT] Registered plugin creator - ::InstanceNormalization_TRT version 1 [TRT] detected model format - caffe (extension '.caffemodel') [TRT] desired precision specified for GPU: FASTEST [TRT] requested fasted precision for device GPU without providing valid calibrator, disabling INT8 [TRT] [MemUsageChange] Init CUDA: CPU +356, GPU +0, now: CPU 390, GPU 5451 (MiB) [TRT] [MemUsageSnapshot] Begin constructing builder kernel library: CPU 390 MiB, GPU 5451 MiB [TRT] [MemUsageSnapshot] End constructing builder kernel library: CPU 494 MiB, GPU 5554 MiB [TRT] native precisions detected for GPU: FP32, FP16, INT8 [TRT] selecting fastest native precision for GPU: FP16 [TRT] could not find engine cache networks/bvlc_googlenet.caffemodel.1.1.8201.GPU.FP16.engine [TRT] cache file invalid, profiling network model on device GPU [TRT] [MemUsageChange] Init CUDA: CPU +0, GPU +0, now: CPU 390, GPU 5554 (MiB) [TRT] [MemUsageSnapshot] Begin constructing builder kernel library: CPU 390 MiB, GPU 5554 MiB [TRT] [MemUsageSnapshot] End constructing builder kernel library: CPU 494 MiB, GPU 5554 MiB [TRT] device GPU, loading networks/googlenet.prototxt networks/bvlc_googlenet.caffemodel [TRT] device GPU, configuring network builder [TRT] device GPU, building FP16: ON [TRT] device GPU, building INT8: ON [TRT] device GPU, workspace size: 33554432 [TRT] device GPU, building CUDA engine (this may take a few minutes the first time a network is loaded) [TRT] Applying generic optimizations to the graph for inference. [TRT] Original: 141 layers [TRT] After dead-layer removal: 141 layers [TRT] Running: FCToConvTransform [TRT] Convert layer type of loss3/classifier from FULLY_CONNECTED to CONVOLUTION [TRT] Running: ShuffleErasure [TRT] Removing shuffle_between_pool5/7x7_s1_and_loss3/classifier [TRT] Applying ScaleNodes fusions. [TRT] After scale fusion: 141 layers [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing conv1/7x7_s2 with conv1/relu_7x7 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing conv2/3x3_reduce with conv2/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing conv2/3x3 with conv2/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/1x1 with inception_3a/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/3x3_reduce with inception_3a/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/3x3 with inception_3a/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/5x5_reduce with inception_3a/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/5x5 with inception_3a/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3a/pool_proj with inception_3a/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/1x1 with inception_3b/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/3x3_reduce with inception_3b/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/3x3 with inception_3b/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/5x5_reduce with inception_3b/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/5x5 with inception_3b/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_3b/pool_proj with inception_3b/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/1x1 with inception_4a/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/3x3_reduce with inception_4a/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/3x3 with inception_4a/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/5x5_reduce with inception_4a/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/5x5 with inception_4a/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4a/pool_proj with inception_4a/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/1x1 with inception_4b/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/3x3_reduce with inception_4b/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/3x3 with inception_4b/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/5x5_reduce with inception_4b/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/5x5 with inception_4b/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4b/pool_proj with inception_4b/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/1x1 with inception_4c/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/3x3_reduce with inception_4c/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/3x3 with inception_4c/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/5x5_reduce with inception_4c/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/5x5 with inception_4c/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4c/pool_proj with inception_4c/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/1x1 with inception_4d/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/3x3_reduce with inception_4d/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/3x3 with inception_4d/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/5x5_reduce with inception_4d/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/5x5 with inception_4d/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4d/pool_proj with inception_4d/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/1x1 with inception_4e/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/3x3_reduce with inception_4e/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/3x3 with inception_4e/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/5x5_reduce with inception_4e/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/5x5 with inception_4e/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_4e/pool_proj with inception_4e/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/1x1 with inception_5a/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/3x3_reduce with inception_5a/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/3x3 with inception_5a/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/5x5_reduce with inception_5a/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/5x5 with inception_5a/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5a/pool_proj with inception_5a/relu_pool_proj [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/1x1 with inception_5b/relu_1x1 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/3x3_reduce with inception_5b/relu_3x3_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/3x3 with inception_5b/relu_3x3 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/5x5_reduce with inception_5b/relu_5x5_reduce [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/5x5 with inception_5b/relu_5x5 [TRT] Running: ConvReluFusion [TRT] ConvReluFusion: Fusing inception_5b/pool_proj with inception_5b/relu_pool_proj [TRT] After vertical fusions: 84 layers [TRT] After dupe layer removal: 84 layers [TRT] After final dead-layer removal: 84 layers [TRT] Merging layers: inception_3a/1x1 + inception_3a/relu_1x1 || inception_3a/3x3_reduce + inception_3a/relu_3x3_reduce || inception_3a/5x5_reduce + inception_3a/relu_5x5_reduce [TRT] Merging layers: inception_3b/1x1 + inception_3b/relu_1x1 || inception_3b/3x3_reduce + inception_3b/relu_3x3_reduce || inception_3b/5x5_reduce + inception_3b/relu_5x5_reduce [TRT] Merging layers: inception_4a/1x1 + inception_4a/relu_1x1 || inception_4a/3x3_reduce + inception_4a/relu_3x3_reduce || inception_4a/5x5_reduce + inception_4a/relu_5x5_reduce [TRT] Merging layers: inception_4b/1x1 + inception_4b/relu_1x1 || inception_4b/3x3_reduce + inception_4b/relu_3x3_reduce || inception_4b/5x5_reduce + inception_4b/relu_5x5_reduce [TRT] Merging layers: inception_4c/1x1 + inception_4c/relu_1x1 || inception_4c/3x3_reduce + inception_4c/relu_3x3_reduce || inception_4c/5x5_reduce + inception_4c/relu_5x5_reduce [TRT] Merging layers: inception_4d/5x5_reduce + inception_4d/relu_5x5_reduce || inception_4d/1x1 + inception_4d/relu_1x1 || inception_4d/3x3_reduce + inception_4d/relu_3x3_reduce [TRT] Merging layers: inception_4e/1x1 + inception_4e/relu_1x1 || inception_4e/3x3_reduce + inception_4e/relu_3x3_reduce || inception_4e/5x5_reduce + inception_4e/relu_5x5_reduce [TRT] Merging layers: inception_5a/1x1 + inception_5a/relu_1x1 || inception_5a/3x3_reduce + inception_5a/relu_3x3_reduce || inception_5a/5x5_reduce + inception_5a/relu_5x5_reduce [TRT] Merging layers: inception_5b/1x1 + inception_5b/relu_1x1 || inception_5b/3x3_reduce + inception_5b/relu_3x3_reduce || inception_5b/5x5_reduce + inception_5b/relu_5x5_reduce [TRT] After tensor merging: 66 layers [TRT] Eliminating concatenation inception_5b/output [TRT] Generating copy for inception_5b/1x1 + inception_5b/relu_1x1 || inception_5b/3x3_reduce + inception_5b/relu_3x3_reduce || inception_5b/5x5_reduce + inception_5b/relu_5x5_reduce to inception_5b/output because input is not movable. [TRT] Retargeting inception_5b/3x3 to inception_5b/output [TRT] Retargeting inception_5b/5x5 to inception_5b/output [TRT] Retargeting inception_5b/pool_proj to inception_5b/output [TRT] Eliminating concatenation inception_5a/output [TRT] Generating copy for inception_5a/1x1 + inception_5a/relu_1x1 || inception_5a/3x3_reduce + inception_5a/relu_3x3_reduce || inception_5a/5x5_reduce + inception_5a/relu_5x5_reduce to inception_5a/output because input is not movable. [TRT] Retargeting inception_5a/3x3 to inception_5a/output [TRT] Retargeting inception_5a/5x5 to inception_5a/output [TRT] Retargeting inception_5a/pool_proj to inception_5a/output [TRT] Eliminating concatenation inception_4e/output [TRT] Generating copy for inception_4e/1x1 + inception_4e/relu_1x1 || inception_4e/3x3_reduce + inception_4e/relu_3x3_reduce || inception_4e/5x5_reduce + inception_4e/relu_5x5_reduce to inception_4e/output because input is not movable. [TRT] Retargeting inception_4e/3x3 to inception_4e/output [TRT] Retargeting inception_4e/5x5 to inception_4e/output [TRT] Retargeting inception_4e/pool_proj to inception_4e/output [TRT] Eliminating concatenation inception_4d/output [TRT] Generating copy for inception_4d/5x5_reduce + inception_4d/relu_5x5_reduce || inception_4d/1x1 + inception_4d/relu_1x1 || inception_4d/3x3_reduce + inception_4d/relu_3x3_reduce to inception_4d/output because input is not movable. [TRT] Retargeting inception_4d/3x3 to inception_4d/output [TRT] Retargeting inception_4d/5x5 to inception_4d/output [TRT] Retargeting inception_4d/pool_proj to inception_4d/output [TRT] Eliminating concatenation inception_4c/output [TRT] Generating copy for inception_4c/1x1 + inception_4c/relu_1x1 || inception_4c/3x3_reduce + inception_4c/relu_3x3_reduce || inception_4c/5x5_reduce + inception_4c/relu_5x5_reduce to inception_4c/output because input is not movable. [TRT] Retargeting inception_4c/3x3 to inception_4c/output [TRT] Retargeting inception_4c/5x5 to inception_4c/output [TRT] Retargeting inception_4c/pool_proj to inception_4c/output [TRT] Eliminating concatenation inception_4b/output [TRT] Generating copy for inception_4b/1x1 + inception_4b/relu_1x1 || inception_4b/3x3_reduce + inception_4b/relu_3x3_reduce || inception_4b/5x5_reduce + inception_4b/relu_5x5_reduce to inception_4b/output because input is not movable. [TRT] Retargeting inception_4b/3x3 to inception_4b/output [TRT] Retargeting inception_4b/5x5 to inception_4b/output [TRT] Retargeting inception_4b/pool_proj to inception_4b/output [TRT] Eliminating concatenation inception_4a/output [TRT] Generating copy for inception_4a/1x1 + inception_4a/relu_1x1 || inception_4a/3x3_reduce + inception_4a/relu_3x3_reduce || inception_4a/5x5_reduce + inception_4a/relu_5x5_reduce to inception_4a/output because input is not movable. [TRT] Retargeting inception_4a/3x3 to inception_4a/output [TRT] Retargeting inception_4a/5x5 to inception_4a/output [TRT] Retargeting inception_4a/pool_proj to inception_4a/output [TRT] Eliminating concatenation inception_3b/output [TRT] Generating copy for inception_3b/1x1 + inception_3b/relu_1x1 || inception_3b/3x3_reduce + inception_3b/relu_3x3_reduce || inception_3b/5x5_reduce + inception_3b/relu_5x5_reduce to inception_3b/output because input is not movable. [TRT] Retargeting inception_3b/3x3 to inception_3b/output [TRT] Retargeting inception_3b/5x5 to inception_3b/output [TRT] Retargeting inception_3b/pool_proj to inception_3b/output [TRT] Eliminating concatenation inception_3a/output [TRT] Generating copy for inception_3a/1x1 + inception_3a/relu_1x1 || inception_3a/3x3_reduce + inception_3a/relu_3x3_reduce || inception_3a/5x5_reduce + inception_3a/relu_5x5_reduce to inception_3a/output because input is not movable. [TRT] Retargeting inception_3a/3x3 to inception_3a/output [TRT] Retargeting inception_3a/5x5 to inception_3a/output [TRT] Retargeting inception_3a/pool_proj to inception_3a/output [TRT] After concat removal: 66 layers [TRT] Graph construction and optimization completed in 0.0624339 seconds. [TRT] ---------- Layers Running on DLA ---------- [TRT] ---------- Layers Running on GPU ---------- [TRT] [GpuLayer] conv1/7x7_s2 + conv1/relu_7x7 [TRT] [GpuLayer] pool1/3x3_s2 [TRT] [GpuLayer] pool1/norm1 [TRT] [GpuLayer] conv2/3x3_reduce + conv2/relu_3x3_reduce [TRT] [GpuLayer] conv2/3x3 + conv2/relu_3x3 [TRT] [GpuLayer] conv2/norm2 [TRT] [GpuLayer] pool2/3x3_s2 [TRT] [GpuLayer] inception_3a/1x1 + inception_3a/relu_1x1 || inception_3a/3x3_reduce + inception_3a/relu_3x3_reduce || inception_3a/5x5_reduce + inception_3a/relu_5x5_reduce [TRT] [GpuLayer] inception_3a/3x3 + inception_3a/relu_3x3 [TRT] [GpuLayer] inception_3a/5x5 + inception_3a/relu_5x5 [TRT] [GpuLayer] inception_3a/pool [TRT] [GpuLayer] inception_3a/pool_proj + inception_3a/relu_pool_proj [TRT] [GpuLayer] inception_3a/1x1 copy [TRT] [GpuLayer] inception_3b/1x1 + inception_3b/relu_1x1 || inception_3b/3x3_reduce + inception_3b/relu_3x3_reduce || inception_3b/5x5_reduce + inception_3b/relu_5x5_reduce [TRT] [GpuLayer] inception_3b/3x3 + inception_3b/relu_3x3 [TRT] [GpuLayer] inception_3b/5x5 + inception_3b/relu_5x5 [TRT] [GpuLayer] inception_3b/pool [TRT] [GpuLayer] inception_3b/pool_proj + inception_3b/relu_pool_proj [TRT] [GpuLayer] inception_3b/1x1 copy [TRT] [GpuLayer] pool3/3x3_s2 [TRT] [GpuLayer] inception_4a/1x1 + inception_4a/relu_1x1 || inception_4a/3x3_reduce + inception_4a/relu_3x3_reduce || inception_4a/5x5_reduce + inception_4a/relu_5x5_reduce [TRT] [GpuLayer] inception_4a/3x3 + inception_4a/relu_3x3 [TRT] [GpuLayer] inception_4a/5x5 + inception_4a/relu_5x5 [TRT] [GpuLayer] inception_4a/pool [TRT] [GpuLayer] inception_4a/pool_proj + inception_4a/relu_pool_proj [TRT] [GpuLayer] inception_4a/1x1 copy [TRT] [GpuLayer] inception_4b/1x1 + inception_4b/relu_1x1 || inception_4b/3x3_reduce + inception_4b/relu_3x3_reduce || inception_4b/5x5_reduce + inception_4b/relu_5x5_reduce [TRT] [GpuLayer] inception_4b/3x3 + inception_4b/relu_3x3 [TRT] [GpuLayer] inception_4b/5x5 + inception_4b/relu_5x5 [TRT] [GpuLayer] inception_4b/pool [TRT] [GpuLayer] inception_4b/pool_proj + inception_4b/relu_pool_proj [TRT] [GpuLayer] inception_4b/1x1 copy [TRT] [GpuLayer] inception_4c/1x1 + inception_4c/relu_1x1 || inception_4c/3x3_reduce + inception_4c/relu_3x3_reduce || inception_4c/5x5_reduce + inception_4c/relu_5x5_reduce [TRT] [GpuLayer] inception_4c/3x3 + inception_4c/relu_3x3 [TRT] [GpuLayer] inception_4c/5x5 + inception_4c/relu_5x5 [TRT] [GpuLayer] inception_4c/pool [TRT] [GpuLayer] inception_4c/pool_proj + inception_4c/relu_pool_proj [TRT] [GpuLayer] inception_4c/1x1 copy [TRT] [GpuLayer] inception_4d/5x5_reduce + inception_4d/relu_5x5_reduce || inception_4d/1x1 + inception_4d/relu_1x1 || inception_4d/3x3_reduce + inception_4d/relu_3x3_reduce [TRT] [GpuLayer] inception_4d/3x3 + inception_4d/relu_3x3 [TRT] [GpuLayer] inception_4d/5x5 + inception_4d/relu_5x5 [TRT] [GpuLayer] inception_4d/pool [TRT] [GpuLayer] inception_4d/pool_proj + inception_4d/relu_pool_proj [TRT] [GpuLayer] inception_4d/1x1 copy [TRT] [GpuLayer] inception_4e/1x1 + inception_4e/relu_1x1 || inception_4e/3x3_reduce + inception_4e/relu_3x3_reduce || inception_4e/5x5_reduce + inception_4e/relu_5x5_reduce [TRT] [GpuLayer] inception_4e/3x3 + inception_4e/relu_3x3 [TRT] [GpuLayer] inception_4e/5x5 + inception_4e/relu_5x5 [TRT] [GpuLayer] inception_4e/pool [TRT] [GpuLayer] inception_4e/pool_proj + inception_4e/relu_pool_proj [TRT] [GpuLayer] inception_4e/1x1 copy [TRT] [GpuLayer] pool4/3x3_s2 [TRT] [GpuLayer] inception_5a/1x1 + inception_5a/relu_1x1 || inception_5a/3x3_reduce + inception_5a/relu_3x3_reduce || inception_5a/5x5_reduce + inception_5a/relu_5x5_reduce [TRT] [GpuLayer] inception_5a/3x3 + inception_5a/relu_3x3 [TRT] [GpuLayer] inception_5a/5x5 + inception_5a/relu_5x5 [TRT] [GpuLayer] inception_5a/pool [TRT] [GpuLayer] inception_5a/pool_proj + inception_5a/relu_pool_proj [TRT] [GpuLayer] inception_5a/1x1 copy [TRT] [GpuLayer] inception_5b/1x1 + inception_5b/relu_1x1 || inception_5b/3x3_reduce + inception_5b/relu_3x3_reduce || inception_5b/5x5_reduce + inception_5b/relu_5x5_reduce [TRT] [GpuLayer] inception_5b/3x3 + inception_5b/relu_3x3 [TRT] [GpuLayer] inception_5b/5x5 + inception_5b/relu_5x5 [TRT] [GpuLayer] inception_5b/pool [TRT] [GpuLayer] inception_5b/pool_proj + inception_5b/relu_pool_proj [TRT] [GpuLayer] inception_5b/1x1 copy [TRT] [GpuLayer] pool5/7x7_s1 [TRT] [GpuLayer] loss3/classifier [TRT] [GpuLayer] prob [TRT] Using cublas as a tactic source [TRT] [MemUsageChange] Init cuBLAS/cuBLASLt: CPU +226, GPU +229, now: CPU 790, GPU 5852 (MiB) [TRT] Using cuDNN as a tactic source [TRT] [MemUsageChange] Init cuDNN: CPU +307, GPU +307, now: CPU 1097, GPU 6159 (MiB) [TRT] Local timing cache in use. Profiling results in this builder pass will not be stored. [TRT] Constructing optimization profile number 0 [1/1]. [TRT] Reserving memory for activation tensors. Host: 0 bytes Device: 606112 bytes [TRT] =============== Computing reformatting costs [TRT] *************** Autotuning Reformat: Float(150528,50176,224,1) -> Float(150528,1,672,3) *************** [TRT] --------------- Timing Runner: Optimizer Reformat(data -> ) (Reformat) [TRT] 2: [utils.cpp::checkMemLimit::380] Error Code 2: Internal Error (Assertion upperBound != 0 failed. Unknown embedded device detected. Please update the table with the entry: {{1794, 6, 16}, 12660},) [TRT] device GPU, failed to build CUDA engine [TRT] device GPU, failed to load networks/bvlc_googlenet.caffemodel [TRT] failed to load networks/bvlc_googlenet.caffemodel [TRT] imageNet -- failed to initialize. imagenet: failed to initialize imageNet