Hello, I have an application which grabs images from a camera
Each image will be processed by VPI CUDA and output to gstreamer
Less frequently, an image is to be analysed by TensorRT yolo (inference time longer than several frames from camera) - with the result informing the parameters of transform to be conducted in the previous task
What strategy should I take for this ? The problem is that the images should be output at same framerate as input and I cannot wait for the inference to finish if its inline
I will only be using the warp perspective VPI, I am thinking maybe this does not need CUDA and can be done on the CPU if it can be done in realtime? Deepstream looks very interesting, are you suggesting it is flexible enough to handle asynchronous processes that occasionally inform the other? EG - an inference working at 1 fps can update an input parameter to a warp perspective which is happening at 30fps?
Warping is a relatively complicated task so it’s recommended to give it a try.
In some cases, the CPU backend can be 10 times slower than the CUDA backend.
Deepstream can do inference periodically and generate the intermediate data with the tracker (detection case).
It will be a good direction to try if the above testing cannot reach 30fps.