Hi guys
I tried to build a TensorRT engine for DLA from the ONNX model with trtexec
But it failed with the following errors
Is there any solution to solve it?
I want to know if this is because I used a custom kernel for torch.nn.functional.conv2d in PyTorch
Machine:
Jetson AGX Orin 64G with JetPack 6.0
command:trtexec --onnx=/home/cast/6_24/small_6_24_1024_rb_round_sim.onnx --saveEngine=/home/cast/6_25/small_6_24_1024_rb_round_sim.trt --shapes=input:32x1x1024x1024 --builderOptimizationLevel=5 --int8 --fp16 --noTF32 --useDLACore=0 --explicitBatch --inputIOFormats=fp16:chw --outputIOFormats=fp16:chw --verbose >dla_verbose_626.txt
Error:
trtexec --onnx=/home/cast/6_24/small_6_24_1024_rb_round_sim.onnx --saveEngine=/home/cast/6_25/small_6_24_1024_rb_round_sim.trt --shapes=input:32x1x1024x1024 --builderOptimizationLevel=5 --int8 --fp16 --noTF32 --useDLACore=0 --explicitBatch --inputIOFormats=fp16:chw --outputIOFormats=fp16:chw --verbose >dla_verbose_626.txt
[06/26/2024-11:15:00] [W] --explicitBatch flag has been deprecated and has no effect!
[06/26/2024-11:15:00] [W] Explicit batch dim is automatically enabled if input model is ONNX or if dynamic shapes are provided when the engine is built.
[06/26/2024-11:15:07] [W] [TRT] onnx2trt_utils.cpp:372: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
[06/26/2024-11:15:07] [E] Error[2]: [network.cpp::operator()::2789] Error Code 2: Internal Error (Assertion allowGPUFallback failed. Layer ‘/Constant_output_0’ is not supported on DLA but GPU fallback is not enabled.)
[06/26/2024-11:15:07] [E] Error[4]: [network.cpp::validate::2901] Error Code 4: Internal Error (DLA validation failed)
[06/26/2024-11:15:07] [E] Engine could not be created from network
[06/26/2024-11:15:07] [E] Building engine failed
[06/26/2024-11:15:07] [E] Failed to create engine from model or file.
[06/26/2024-11:15:07] [E] Engine set up failed