Please provide complete information as applicable to your setup.
• Hardware Platform (Jetson / GPU)
GPU • DeepStream Version
6.1.1 • JetPack Version (valid for Jetson only) • TensorRT Version • NVIDIA GPU Driver Version (valid for GPU only) • Issue Type( questions, new requirements, bugs)
questions • How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing) • Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description)
Hello,
How should the mask output data from PeopleSegNet (v2.0.2) be interpreted?
Each detected object has ‘mask_params’ metadata attached as below:
Could you please elaborate on the format of the output and how to convert it to a pixel mask to overlay on an image?
I’ve attached the float[3136] array from above as a .txt - How can I convert and scale the mask to fit the image?
The detection box is:
left: 251.45857
top: 270.86142
width: 171.64272
height: 243.53386 sample.txt (31.0 KB)
The NGC provides no real information on how to interpret the outputs except for “Category label (person), bounding-box coordinates and segmentation mask for each detected person in the input image.”
Could you elaborate on how to convert the float[3126] vector from mask_params->data into a pixel mask I can overlay on the output video?
Sorry, my fault. I misread the structure of your remarks. These datas you marked are float numbers output from the inference. You can convert it to a argb32 pixel format. This part of the code is not open source, so I suggest you refer to the API of the conversion:nvds_mask_utils_resize_to_binary_argb32 from nvds_mask_utils.h