Please provide complete information as applicable to your setup.
• Hardware Platform (Jetson / GPU):v100
• DeepStream Version: 5.0.0
• JetPack Version (valid for Jetson only)
• TensorRT Version :from docker images
• NVIDIA GPU Driver Version (valid for GPU only):450.88
• Issue Type( questions, new requirements, bugs)questions
• How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing) :deepstream_test_1.py run error
• Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description) :python3 deepstream_test_1.py /opt/nvidia/deepstream/deepstream-5.0/samples/streams/sample_720p.h264
I run command deepstream-app --version-all
,get print:
deepstream-app version 5.0.0
DeepStreamSDK 5.0.0
CUDA Driver Version: 11.0
CUDA Runtime Version: 10.2
TensorRT Version: 7.0
cuDNN Version: 7.6
libNVWarp360 Version: 2.0.1d3
hello, I’m using docker container for dGPU - nvcr.io/nvidia/deepstream:5.0.1-20.09-triton and downloaded latest DeepStream 5.0.1 Python apps inside this container,when I run python3 deepstream_test_1.py /opt/nvidia/deepstream/deepstream-5.0/samples/streams/sample_720p.h264
,error occurred,can you help me? the error is :
Creating Pipeline
Creating Source
Creating H264Parser
Creating Decoder
Creating EGLSink
Playing file /opt/nvidia/deepstream/deepstream-5.0/samples/streams/sample_720p.h264
Adding elements to PipelineLinking elements in the Pipeline
Starting pipeline
WARNING: …/nvdsinfer/nvdsinfer_func_utils.cpp:36 [TRT]: Current optimization profile is: 0. Please ensure there are no enqueued operations pending in this context prior to switching profiles
0:00:02.882016620 16130 0xff5fd00 INFO nvinfer gstnvinfer.cpp:619:gst_nvinfer_logger: NvDsInferContext[UID 1]: Info from NvDsInferContextImpl::deserializeEngineAndBackend() <nvdsinfer_context_impl.cpp:1701> [UID = 1]: deserialized trt engine from :/opt/nvidia/deepstream/deepstream-5.0/samples/models/Primary_Detector/resnet10.caffemodel_b1_gpu0_int8.engine
INFO: …/nvdsinfer/nvdsinfer_model_builder.cpp:685 [Implicit Engine Info]: layers num: 3
0 INPUT kFLOAT input_1 3x368x640
1 OUTPUT kFLOAT conv2d_bbox 16x23x40
2 OUTPUT kFLOAT conv2d_cov/Sigmoid 4x23x400:00:02.882142504 16130 0xff5fd00 INFO nvinfer gstnvinfer.cpp:619:gst_nvinfer_logger: NvDsInferContext[UID 1]: Info from NvDsInferContextImpl::generateBackendContext() <nvdsinfer_context_impl.cpp:1805> [UID = 1]: Use deserialized engine model: /opt/nvidia/deepstream/deepstream-5.0/samples/models/Primary_Detector/resnet10.caffemodel_b1_gpu0_int8.engine
0:00:02.986582772 16130 0xff5fd00 INFO nvinfer gstnvinfer_impl.cpp:313:notifyLoadModelStatus: [UID 1]: Load new model:dstest1_pgie_config.txt sucessfully
Frame Number=0 Number of Objects=6 Vehicle_count=4 Person_count=2
0:00:03.258218015 16130 0x2c29230 WARN nvinfer gstnvinfer.cpp:1975:gst_nvinfer_output_loop: error: Internal data stream error.
0:00:03.258233862 16130 0x2c29230 WARN nvinfer gstnvinfer.cpp:1975:gst_nvinfer_output_loop: error: streaming stopped, reason not-negotiated (-4)
Frame Number=1 Number of Objects=6 Vehicle_count=4 Person_count=2
Error: gst-stream-error-quark: Internal data stream error. (1): gstnvinfer.cpp(1975): gst_nvinfer_output_loop (): /GstPipeline:pipeline0/GstNvInfer:primary-inference:
streaming stopped, reason not-negotiated (-4)
thank you very much!