Hello, We planned to use DINO model instead of Yolor model and since the inference speed is high for fp16. We planned on convert to int8 mode but found that it is not available for DINO. Will int8 mode for DINO will support in future, and if not what do you suggest to improve the inference speed.
I will sync internally for the feature request for the int8. For improving the inference speed, you can select a smaller backbone. For example, resnet_50, fan_tiny, gcvit_xxtiny.
Hi, Have few doubts, You have stated int8 is not supported for DINO, can I know what then here DINO | NVIDIA NGC what does the int8 stated here refer to. Any assistance provided regarding int8 mode is highly appreciated.
There is no update from you for a period, assuming this is not an issue anymore. Hence we are closing this topic. If need further support, please open a new one. Thanks