ONNX to tensorrt model file not support int8 yet, in ONNX InceptionV1 TensorRT 5.0.2 V100-16G INT8 | NVIDIA NGC, i found a inceptionv1 inference tensorrt model, but it’s max batchsize is 1, where can i get a bigger batchsize model, a guide with how to do the model transform will be appreciated.
Related topics
Topic | Replies | Views | Activity | |
---|---|---|---|---|
TensorRT INT8 conversion from an ONNX model | 4 | 5509 | July 29, 2024 | |
Model does not get Int8 layers | 4 | 525 | September 19, 2022 | |
TensorRT INT8 inference accuracy | 2 | 498 | May 9, 2022 | |
ONNX batchsize setting and buffer.h assert error | 3 | 1174 | March 23, 2021 | |
TensorRT5.0.x INT8 for Onnx? | 2 | 963 | October 12, 2021 | |
Batch inference on tensorrt | 4 | 427 | February 15, 2021 | |
Converting ONNX model to INT8 | 3 | 3282 | April 9, 2021 | |
Tensorrt ONNX Inference result mismatch with tensorflow 2.0 Inference | 1 | 772 | February 18, 2020 | |
Tensorrt Conversion from ONNX for keras-ocr models fails because of int32 input in intermediate layers | 1 | 750 | March 8, 2023 | |
Uff to trt8 | 6 | 866 | January 19, 2022 |