Please provide complete information as applicable to your setup.
• Hardware Platform (Jetson / GPU)
GPU • DeepStream Version
6.1.1 • JetPack Version (valid for Jetson only) • TensorRT Version • NVIDIA GPU Driver Version (valid for GPU only) • Issue Type( questions, new requirements, bugs)
questions • How to reproduce the issue ? (This is for bugs. Including which sample app is using, the configuration files content, the command line used and other details for reproducing) • Requirement details( This is for new requirement. Including the module name-for which plugin or for which sample application, the function description)
How should the mask output data from PeopleSegNet (v2.0.2) be interpreted?
Each detected object has ‘mask_params’ metadata attached as below:
Could you please elaborate on the format of the output and how to convert it to a pixel mask to overlay on an image?
I’ve attached the float array from above as a .txt - How can I convert and scale the mask to fit the image?
The detection box is:
height: 243.53386 sample.txt (31.0 KB)
Sorry, my fault. I misread the structure of your remarks. These datas you marked are float numbers output from the inference. You can convert it to a argb32 pixel format. This part of the code is not open source, so I suggest you refer to the API of the conversion:nvds_mask_utils_resize_to_binary_argb32 from nvds_mask_utils.h