Hi there,
I’m trying to modify the yolov7 model to inference on DLA core of Orin/Xavier NX devices.
But my Add operations’ outputs are fellback to GPU.
As it referred in the document:
DLA does support the element-wise ops with same shape inputs. My 2 inputs are at the same shape of [1, 2, H, W] then why the outputs are fellback to GPU?
The terminal outputs are provided below.
Thanks.
log.txt (379.7 KB)
I run this command and log the result to the attached file: trtexec --int8 --inputIOFormats=int8:dla_hwc4 --outputIOFormats=int8:chw32 --onnx=best_mod.onnx --useDLACore=0 --allowGPUFallback --verbose
More infomation about the fellback layers which is returned in the onnx conversion:
*%/model.77/Constant_18_output_0 : Float(1, 2, 80, 80, strides=[12800, 6400, 80, 1], requires_grad=0, device=cpu) = onnx::Constantvalue=, onnx_name=“/model.77/Constant_18”, scope: models.yolo.Model::/models.yolo.IDetect::model.77 *
Yes, one of my inputs is constant. When I search for the fallback layers on netron, it points out the constant input of the add operations. Then does the DLA not support addition with constant input?
I have just realized that DLA does not support constant layers. The add operations are supported only if the 2 inputs are not constant.
Btw, thanks for your reply!