Therefore, using DLA may not get higher performance.
How could we get higher inference performance on NX?
The document of NX shows Tiny Yolo V3 performance is around 562 fps.
However, we got the test result below.
Yolov3 with interval=2 got **PERF: 57.03 (55.15)
Yolov3 with interval=0 got **PERF: 25~26
Are the test results normal?
or anything we can do to improve it?
Plus, is there a way to enable two DLA at the same time? How?
Does enabled two DLA at the same time help on inference performance or not?
Sorry that there are some unclear statement in my previous reply.
It won’t get you a higher performance if inference with only DLA.
Since the target for DLA is power saving and offload GPU usage.
However, for XavierNX, you can try to use 2DLA and GPU together.
This will get you the maximal performance of XavierNX.
We also use 2DLA together with GPU to get the 562 performance.
To reproduce the benchmark result, you can use our script here directly:
Yes, I did follow the website you provided.
Here are the steps I did. git clone https://github.com/NVIDIA-AI-IOT/jetson_benchmarks.git cd jetson_benchmarks mkdir models sudo sh install_requirements.sh python3 utils/download_models.py --all --csv_file_path <path-to>/benchmark_csv/nx-benchmarks.csv --save_dir <absolute-path-to-downloaded-models>
You need to specify the complete dir paths to run benchmarks (but not needed for downloading models).
For example (if you name is “XXX”):
python3 utils/download_models.py --all --csv_file_path benchmark_csv/nx-benchmarks.csv --save_dir models
sudo python3 benchmark.py --all --csv_file_path /home/XXX/jetson_benchmarks/benchmark_csv/nx-benchmarks.csv --model_dir /home/XXX/jetson_benchmarks/models
Thank you for your remind.
(somehow, I did not get Pelepicier’s update.)
I could get the test result normally by following Pelepicier’s suggestions.
Hi Pelepicier,
Thank you so much for your help.
The suggestion you provided works.