Hello, I have an application which grabs images from a camera
Each image will be processed by VPI CUDA and output to gstreamer
Less frequently, an image is to be analysed by TensorRT yolo (inference time longer than several frames from camera) - with the result informing the parameters of transform to be conducted in the previous task
What strategy should I take for this ? The problem is that the images should be output at same framerate as input and I cannot wait for the inference to finish if its inline
I will only be using the warp perspective VPI, I am thinking maybe this does not need CUDA and can be done on the CPU if it can be done in realtime? Deepstream looks very interesting, are you suggesting it is flexible enough to handle asynchronous processes that occasionally inform the other? EG - an inference working at 1 fps can update an input parameter to a warp perspective which is happening at 30fps?