with the reference of [url]https://devblogs.nvidia.com/int8-inference-autonomous-vehicles-tensorrt/[/url], we managed to use the calibrator from python interface to convert the MNIST caffemodel to INT8 inference engine. we can dump the weight from the original caffemodel in FP32 through the protobuf utility. is there a way that we can dump the INT8 engine weights?
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
Int8 weight extraction | 2 | 521 | April 27, 2020 | |
Can TensorRT 7.1.3 convert an INT8 pytorch QAT model to engine? | 3 | 706 | April 21, 2022 | |
Data process about TensorRT INT8 and FP16 inference Engine | 4 | 1994 | October 18, 2021 | |
How to create a new model? | 1 | 517 | June 27, 2022 | |
Building INT8 engine fails: ../builder/cudnnBuilderWeightConverters.cpp:97:... | 2 | 795 | October 12, 2021 | |
Does TensorRT Python API support uff model int8 calibration | 1 | 976 | July 11, 2019 | |
Data inferencing to INT8U quantized model | 2 | 407 | October 12, 2021 | |
TensorRT 8-bit Quantization questions | 7 | 4796 | April 26, 2018 | |
How to use DLA + INT8 + I/O reformatting? | 1 | 598 | June 13, 2023 | |
pre-quantized models on Jetson AGX Xavier | 10 | 940 | October 18, 2021 |