NVIDIA Developer Forums
Post quantization aware training is slower than fp16 and post quantization
AI & Data Science
Deep Learning (Training & Inference)
TensorRT
OnePieceOfDeepLearning
September 27, 2021, 3:46am
7
int8 is supported in jetson nx. we got very great speed with PTQ.
QAT int8 TRT engine slower than fp16
show post in topic