I have achieved nvinferserver plugin function in deepstream with deepstream-app, such as deepstream-app -c source1_primary_detector_nano.txt, and can run it succeed.
However, I want deploy triton server in dGPU and deploy deepstream in jetson, and all pipeline get infer result from triton server, including preprocess and postprocess.
And another solution is start different pipeline and share one nvinferserver in jetson or dGPU, as I observed that there are different server when I start different pipeline, it cost lots resources.
Can you tell me how to achieve it? Or whether it is possible. Thanks!
Hi,
You can always integrate Triton Client in Deepstream,
Comes as python/cpp ,
Can integrate cpp client in Deepstream dsexample plugin, or python client in python probe (Deepstream python bindings)
I am looking for similar kind of solution and I have referred triton client example with grpc but those are related to single images only, How can do it for videos? Can you please share any reference example which we can follow to achieve same?
Do I need to consider video as series of images and need call triton infer server with every frame?
How Do you recommend if I need to Deeptream pipeline on same host as triton server?
Hi,
Yes, efficient decode via deepstream and send each image to inference via cuda shared mem triton api for best performance on same machine / grpc for remote server.
Does Triton Server supports multi stage inference and Tracker in single request to Triton Inference Server or Do we need to make separate request for each stage inference? Does Triton Inference Server help tracking object if we have pipeline like below.