Hi I am a beginner in this field so I require some help. I have a NVIDIA AGX Orin 64 GP Developer Kit and want to implement Realtime Facial Recognition on it. I read the documentation (Developer Guide) but have some queries regarding model conversion
Can I convert any model from .onnx to tensorRT engine (eg. arcFace, Resnet100, etc.) using trtexec ??
If my TensorRT is utilizing the GPU properly is the above conversion the most efficient way to run the model ?
Is the usage of –useDLACore=0 while making the tensorRT engine enough to utilize DLA or do I have to menton anything when running these models on DeepStream ?
1. In general yes, but if your model contains a special layer, TensorRT might not have a corresponding implementation for it. 2. Yes, efficient in both performance and memory. 3. You can put the model on DLA with --useDLACore=[ID] when inferring with trtexec.
For Deepstream, please find below the topic to modify the config to run on DLA.
Thanks a lot @AastaLLL How do I verify whether my .engine file of the model I generated is using DLA.
The execution was successful and I passed --useDLACore=0 --allowGPUFallback as arguements when I converted my onnx file to engine. Just for your reference I am using arcface_r100_v1.onnx and converting that to engine. I later plan to use it in Deepstream.