Description
I’m trying to inference Yolov5 with TensorRT on Jetson Nano 4GB, However, the result is quite weird since using original ‘yolov5s.pt’, the inference speed is faster (~120ms) than when using ‘yolov5s.engine’ generated from the producer export.py (~140ms).
Environment
TensorRT Version: TensorRT 8.0.1
GPU Type: Jetson Nano GPU
CUDA Version: CUDA 10.2
CUDNN Version: CUDNN 8.2.1
Operating System + Version: Ubuntu 18.04
Python Version (if applicable): Python 3.6.9
PyTorch Version (if applicable): Pytorch v1.10.0
Relevant Files
Here is the link to my video (1920x1080): https://drive.google.com/file/d/1stEqgGOWqcBFJHmpkWR_-KzABg8k0HLJ/view?usp=sharing
Steps To Reproduce
First, from the original yolov5s.pt, I use this command line to produce TensorRT yolov5s.engine file:
python3 export.py --weights yolov5s.pt --include engine --imgsz 640 640 --device 0
Then, I started inference with this .engine file:
python3 detect.py --weights yolov5s.engine --imgsz 640 640 --device 0 --source ./data/images/1080p.mp4
The result was about ~140ms inferencing on each frames. [link_image]
But when comparing to the original path file with:
python3 detect.py --weights yolov5s.pt --imgsz 640 640 --device 0 --source ./data/images/1080p.mp4
The result was about ~120ms inferencing on each frames [link_images]
Please help me !!!