Dear support team !
In scope of my project, I’ve used OnnxYolov4 model for development by using TensorRT-126.96.36.199 package on host machine.
No issues wasn’t observed during parsing ONNX model and further steps.
The C++ code has been compiled for aarch64 and launched on targedt Drive AGX platform.
As a result Ive gotten error:
[06/01/2021-18:36:05] [W] [TRT] ModelImporter.cpp:140: No importer registered for op: Equal. Attempting to import as plugin.
[06/01/2021-18:36:05] [I] [TRT] builtin_op_importers.cpp:2191: Searching for plugin: Equal, plugin_version: 1, plugin_namespace:
While parsing node number 375 [Equal]:
ERROR: builtin_op_importers.cpp:2193 In function importFallbackPluginImporter:
 Assertion failed: creator && “Plugin not found”
FailEqualOpAgx.log (10.9 KB)
If I understood correctly, TensorRT 6.3.1 package on target platform doesn’t support op: Equal .
In other words the library Onnx model parsing on target platform can’t support this operation:
On the other hand, libnvonnxparser.so.7.2.1 already has this support and it can be explained the difference in results.
Please correct me if I’m wrong.
Could you please help me to understand what is the most fastest and correct solution w/o writing a custom plugin would be done:
- DRIVE OS Linux 5.2.0 has TensorRT 6.3.1, does Nvidia have any plans to provide the latest TensorRT package in the future ?
- Is it possible to compile libnvonnxparser.so.7.2.1 locally on the host for aarch64 or directly on target platform
and replace it ?
I appreciate your help in any case.