如何使能DLA

我正在使用NVIDIA AGX Orin 64GB开发者套件推理 VAD.onnx 模型,但是性能很差(约3.5fps),我注意到jetson_release显示 DLA0 DLA1 off,我应该如何激活DLA0,DLA1,为了使用它们?同时,我可以使用什么工具,查看编译出的 VAD.engine 的性能瓶颈点,提高推理效率
I am using the NVIDIA AGX Orin 64GB Developer Kit to infer the VAD.onnx model, but the performance is very poor (approximately 3.5 FPS). I noticed that jetson_release shows DLA0 and DLA1 as off. How should I enable DLA0 and DLA1 to use them? Additionally, what tools can I use to identify performance bottlenecks in the compiled VAD.engine and improve inference efficiency?

VAD.onnx

Hi,

When deploying a model with TensorRT, please use --useDLACore=[ID] to use DLA for inference.

Please find more details in our document below:

Thanks.

我可以怎样分析模型的性能瓶颈呢?
I need to analyze the performance bottlenecks of the model. How should I do it?

Is “OFF” a normal status?

Hi,

Do you want to profile a model that runs on the DLA?
If so, please find below link for a tutorial:

Thanks.

don’t find nvvp or Nvidia Visual Profiler on Jetson AGX Orin Developer Kit

sudo apt install nvidia-jetpack should install them. After installation, you can use nsys to profile your program and nsys-ui to view your report in GUI.

thank you for your reply

This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.