Please provide complete information as applicable to your setup.
• Hardware Platform (Jetson / GPU) GPU
• DeepStream Version 7.0
• JetPack Version (valid for Jetson only)
• TensorRT Version
• NVIDIA GPU Driver Version (valid for GPU only) 535.161.08
• Issue Type( questions, new requirements, bugs) Questions/Bugs?
• How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing)
• Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description)
Hi, I am trying to implement DS with Triton inference server on Yolo V5 in the multiarch docker mentioned in DS 7.0 Quick Start page, but I have encountered a problem that the inference result directly obtained from onnx in triton (throught gRPC protocal) is incorrect while through onnx runtime the model is working well, check the wrong_output.jpg and correct_output.jpg for example.
for reproducing, here are the files I am using for testing:
- Python binding: deepstream_test1_rtsp_in_rtsp_out.py
- official Github repository for postprocessing: yolo deepstream modified to adapt to yolov5 output format (the original YoloV4 decode/parse function) nvdsparsebbox_Yolo.cpp
- Yolo V5s onnx file exported from official pt file (put in my github repository): model.zip (trained with coco128) , labels in labels.txt in github (run in triton without config.pbtxt)
- deepstream config file dstest1_yolov5s.txt in github
- command line used is python3 deepstream_test1_rtsp_in_rtsp_out.py -i {source} -g nvinferserver
I guess the problem lies in preprocess part, where the input data is not well formatted for onnx model (NCWH), but I don’t have any idea on why this is happening cuz the input is supposed to be automatically processes with configuration TENSOR_ORDER_LINEAR?
Hoping to get some suggestions or help, thanks.