Please provide complete information as applicable to your setup.
• Hardware Platform (Jetson / GPU) GPU
• DeepStream Version 7.0
• JetPack Version (valid for Jetson only)
• TensorRT Version 8.6.1.6
• NVIDIA GPU Driver Version (valid for GPU only)
• Issue Type( questions, new requirements, bugs)
• How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing)
• Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description)
I’m running a YoloV8-seg inside deepstream 7.0 and I’ve encountered this error:
ERROR: [TRT]: 1: [runner.cpp::shapeChangeHelper::621] Error Code 1: Myelin (Division by 0 detected in the shape graph. Tensor (Divisor) “sp__mye3” is equal to 0.; )
ERROR: nvdsinfer_backend.cpp:507 Failed to enqueue trt inference batch
ERROR: nvdsinfer_context_impl.cpp:1824 Infer context enqueue buffer failed, nvinfer error:NVDSINFER_TENSORRT_ERROR
0:05:23.176120329 585203 0x55d26850b440 WARN nvinfer gstnvinfer.cpp:1418:gst_nvinfer_input_queue_loop: error: Failed to queue input batch for inferencing
from this:
I understand that you expect this bug to be solved in later versions of tensorrt
- can you confirm that the bug was solved in updated tensorrt?
- If I understand correctly deepstream sdk 7.0 depend on tensorrt 8.6.1.6, so what are my options to address this issue?
- If there is no solution for this right now, is there an option to catch this error and handle it in my python code? currently it crashes the whole app
thanks