Dear Community,
I was trying to convert Yolov11n-seg model into INT8 precision using TRT-CPP for getting better inference speed. I was using this GitHub repository for understanding the process.
I have also looked for INT8 conversion using TRT-CPP in GitHub and found nothing. Majority is using FP16. Though they claim better inference time, I want to use int8 for my use case.
The below are my Specs:
HW:Jetson Orin Nano Dev Kit
JP:6.2 Super Dev
CUDA:12.6
TRT:10.3.0.30
OpenCV with CUDA: 4.12.0-dev
it would be very helpful if you share your insight on this.
I am also attaching the SS of the error I got during INT8 conversion using the above git repo.
Thank you
Regards
Kiranraj K R
