hi AstaLLL:
The problem is not solved,
root@uu-desktop:/opt/nvidia/deepstream/deepstream-4.0/sources/apps/sample_apps/back-to-back-detectors# ./back-to-back-detectors ../../../../samples/streams/test.mp4
Now playing: ../../../../samples/streams/test.mp4
Using winsys: x11
Opening in BLOCKING MODE
Creating LL OSD context new
0:00:00.714180655 3082 0x55a5182670 INFO nvinfer gstnvinfer.cpp:519:gst_nvinfer_logger:<primary-nvinference-engine2> NvDsInferContext[UID 2]:initialize(): Trying to create engine from model files
Loading pre-trained weights...
Loading complete!
Total Number of weights read : 8858734
layer inp_size out_size weightPtr
(1) conv-bn-leaky 3 x 416 x 416 16 x 416 x 416 496
(2) maxpool 16 x 416 x 416 16 x 208 x 208 496
(3) conv-bn-leaky 16 x 208 x 208 32 x 208 x 208 5232
(4) maxpool 32 x 208 x 208 32 x 104 x 104 5232
(5) conv-bn-leaky 32 x 104 x 104 64 x 104 x 104 23920
(6) maxpool 64 x 104 x 104 64 x 52 x 52 23920
(7) conv-bn-leaky 64 x 52 x 52 128 x 52 x 52 98160
(8) maxpool 128 x 52 x 52 128 x 26 x 26 98160
(9) conv-bn-leaky 128 x 26 x 26 256 x 26 x 26 394096
(10) maxpool 256 x 26 x 26 256 x 13 x 13 394096
(11) conv-bn-leaky 256 x 13 x 13 512 x 13 x 13 1575792
(12) maxpool 512 x 13 x 13 512 x 13 x 13 1575792
(13) conv-bn-leaky 512 x 13 x 13 1024 x 13 x 13 6298480
(14) conv-bn-leaky 1024 x 13 x 13 256 x 13 x 13 6561648
(15) conv-bn-leaky 256 x 13 x 13 512 x 13 x 13 7743344
(16) conv-linear 512 x 13 x 13 255 x 13 x 13 7874159
(17) yolo 255 x 13 x 13 255 x 13 x 13 7874159
(18) route - 256 x 13 x 13 7874159
(19) conv-bn-leaky 256 x 13 x 13 128 x 13 x 13 7907439
(20) upsample 128 x 13 x 13 128 x 26 x 26 -
(21) route - 384 x 26 x 26 7907439
(22) conv-bn-leaky 384 x 26 x 26 256 x 26 x 26 8793199
(23) conv-linear 256 x 26 x 26 255 x 26 x 26 8858734
(24) yolo 255 x 26 x 26 255 x 26 x 26 8858734
Output blob names :
yolo_17
yolo_24
Total number of layers: 50
Total number of layers on DLA: 0
Building the TensorRT Engine...
Building complete!
0:00:29.349012887 3082 0x55a5182670 INFO nvinfer gstnvinfer.cpp:519:gst_nvinfer_logger:<primary-nvinference-engine2> NvDsInferContext[UID 2]:generateTRTModel(): Storing the serialized cuda engine to file at /opt/nvidia/deepstream/deepstream-4.0/sources/apps/sample_apps/back-to-back-detectors/model_b1_fp32.engine
Deserialize yoloLayerV3 plugin: yolo_17
Deserialize yoloLayerV3 plugin: yolo_24
0:00:29.808689508 3082 0x55a5182670 INFO nvinfer gstnvinfer.cpp:519:gst_nvinfer_logger:<primary-nvinference-engine1> NvDsInferContext[UID 1]:initialize(): Trying to create engine from model files
0:00:29.808830915 3082 0x55a5182670 WARN nvinfer gstnvinfer.cpp:515:gst_nvinfer_logger:<primary-nvinference-engine1> NvDsInferContext[UID 1]:generateTRTModel(): INT8 not supported by platform. Trying FP16 mode.
0:01:08.540191801 3082 0x55a5182670 INFO nvinfer gstnvinfer.cpp:519:gst_nvinfer_logger:<primary-nvinference-engine1> NvDsInferContext[UID 1]:generateTRTModel(): Storing the serialized cuda engine to file at /opt/nvidia/deepstream/deepstream-4.0/samples/models/Primary_Detector/resnet10.caffemodel_b1_fp16.engine
Running...
NvMMLiteOpen : Block : BlockType = 261
NVMEDIA: Reading vendor.tegra.display-size : status: 6
NvMMLiteBlockCreate : Block : BlockType = 261
Creating LL OSD context new
Frame Number = 0 Vehicle Count = 6 Person Count = 6 Face Count = 0 License Plate Count = 0
0:01:09.468900354 3082 0x55a5179f20 WARN nvinfer gstnvinfer.cpp:1149:convert_batch_and_push_to_input_thread:<primary-nvinference-engine2> error: NvBufSurfTransform failed with error -2 while converting buffer
0:01:09.469822617 3082 0x55a5179f20 WARN nvinfer gstnvinfer.cpp:1830:gst_nvinfer_output_loop:<primary-nvinference-engine1> error: Internal data stream error.
ERROR from element primary-nvinference-engine2: NvBufSurfTransform failed with error -2 while converting buffer
0:01:09.469968504 3082 0x55a5179f20 WARN nvinfer gstnvinfer.cpp:1830:gst_nvinfer_output_loop:<primary-nvinference-engine1> error: streaming stopped, reason error (-5)
Error details: /dvs/git/dirty/git-master_linux/deepstream/sdk/src/gst-plugins/gst-nvinfer/gstnvinfer.cpp(1149): convert_batch_and_push_to_input_thread (): /GstPipeline:pipeline/GstNvInfer:primary-nvinference-engine2
Returned, stopping playback
0:01:09.489323239 3082 0x55a5179f20 WARN nvinfer gstnvinfer.cpp:1149:convert_batch_and_push_to_input_thread:<primary-nvinference-engine2> error: NvBufSurfTransform failed with error -3 while converting buffer
double free or corruption (out)
Aborted (core dumped)
Does the darknet algorithm support this? Back-to-back uses the caffe model.