Hi, I was wondering if Deepstream offers stitching / concatenating / merging multiple images before inference.
I am planning to run an inference on Jetson Nano for two video sources, then im afraid that the FPS would be significantly low if I wanna use PeopleNet as the resolution would be 960*544.
And the plugin StreamMux(muxer) seems to be offering batch inference only, which seems very different from what I am expecting.
So, I wonder:
- If it is able to stitch multiple videos into one frame of image (960&544 for example), and run inference using Deepstream?
- Is it only possible using TensorRT python script with modification? Example link (jetson-inference/detectnet.py at master · dusty-nv/jetson-inference · GitHub)
++) 3. According to this Guideline on page 50, it says “Optical flow vectors are
useful in various use cases such as object detection and tracking, video frame rate upconversion, depth estimation, stitching, and so on.” https://docs.nvidia.com/metropolis/deepstream/4.0/DeepStream_Plugin_Manual.pdf
But, Im not sure how to employ this GST-NVOF
to stitch images before run an inference.
I do appreciate for your suuport in advance.