Description
I am trying to convert an already quantized onnx model to tensorrt!
When I try to parse my quantized onnx network, I get the following error
In node 1 (parseGraph): UNSUPPORTED_NODE: No importer registered for op: QLinearConv
.
In the list of Tensorrt supported onnx operators here https://github.com/onnx/onnx-tensorrt/blob/master/operators.md, I can see that QlinearConv is not supported.
Is there any guideline on how to convert quantized onnx model to trt?
Environment
TensorRT Version: 7.0.0.11
GPU Type: T4
Nvidia Driver Version: 440.33.01
CUDA Version: 10.2
CUDNN Version: 7605
Operating System + Version: Ubuntu 18.04.5 LTS
Python Version (if applicable): 3.6.9
PyTorch Version (if applicable): 1.6.0
Steps To Reproduce
TRT_LOGGER = trt.Logger(trt.Logger.WARNING)
EXPLICIT_BATCH = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH)
with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser(network, TRT_LOGGER) as parser:
builder.max_workspace_size = max_workspace_size
builder.max_batch_size = 1
if not parser.parse(model):
for i in range(parser.num_errors):
error = parser.get_error(i)
print(error)
and the network
is a quantized onnx model.