I need to track items as they are inserted into a bin. To do that, I have a tiny YOLOv3 detector that identifies the end effector. From that, I trace the contours around the item it is holding and fit a bounding box around the item. I want to run inference on the item with DeepStream, not the end effector. Since I have to provide a custom function to return the bounding boxes from the tiny YOLOv3 model (
parse-bbox-func-name), it seems a natural fit to add my custom logic there. To do that, I need the RGB(A) frames for inference. Is there a way to get them from within the function? If not, how may I build such a model with DeepStream?
The function is given a
NvDsInferLayerInfo objects. I’m assuming each of them represents one image frame, and the documentation says it contains a
(void *) field named
buffer that is a
"Pointer to the buffer for the layer data". Does that mean they have references to their associated image frames? If so, how may I get them?