i trained a yolov4 model with 1536*1536, but FP16 runs only 15FPS.
how can i get a simple way to change int8 for more fast?
pls give me some instructions about that.
thank you
i trained a yolov4 model with 1536*1536, but FP16 runs only 15FPS.
how can i get a simple way to change int8 for more fast?
pls give me some instructions about that.
thank you
Hi,
Which framework do you use for inference?
If it is TensorRT, you can create an INT8 engine with the following command directly:
$ /usr/src/tensorrt/bin/trtexec --int8 ...
Thanks.
thanks
not input problem i am sure, and int8.engine can not find object after create from onnx,
my model is yolov4.
does it need a cal_trt.bin file?
thanks
anybody to answer me about yolov4 int8 questions?
thanks
Hi,
Sorry for the late update.
If your app works well with the fp16 model, then the issue should come from the calibration file.
Do you use the official YOLOv4 model?
If yes, you can try the calibration cache from our user:
Thanks.
yes i use official yolov4-tiny, but it is 1536*1536model which be trained,
can i use yolov4-tiny-406-calib_cache directly for int8?
thanks!
Hi,
Since the input size is different, it’s recommended to re-generate a calibration cache.
You can find an example in GitHub as well:
Thanks.
thank you i will study that.
This topic was automatically closed 60 days after the last reply. New replies are no longer allowed.