Yolov5 - NVIDIA DeepStream and NVIDIA Triton Inference Server

Description

Reference: Deploying Models from TensorFlow Model Zoo Using NVIDIA DeepStream and NVIDIA Triton Inference Server | NVIDIA Developer Blog

nms has no object when I use Yolov5 with deepstream and Trition Inference Server. I thought the problem in my config my at the preprocessing step but I have no idea to fix it. Please help.

Environment

TensorRT Version: 7.2.2.3
GPU Type: GTX 2080
Nvidia Driver Version:
CUDA Version: 11.1
CUDNN Version: 8
Baremetal or Container (if container which image + tag): nvcr.io/nvidia/deepstream:5.1-21.02-triton

Relevant Files

Github: GitHub - ultralytics/yolov5: YOLOv5 in PyTorch > ONNX > CoreML > TFLite
I modified the input shape of model after convert to onnx

model_yolov5s.onnx (27.7 MB)
config.pbtxt (510 Bytes)
labels.txt (620 Bytes)
deepstream_app_config_yolov5.txt (3.7 KB)
config_infer_primary_yolov5.txt (1.2 KB)

Steps To Reproduce

And nms has no object.