Multi-batch infer results with the same input are different and only the first batch is wrong when running on the DLA

JetPack Version : 4.4

The one-batch model running on the DLA in fp16 always comes out the wrong result, but the multi-batch infer result is different with the same input and only the first batch is wrong.
The model using one or more batches running normally on GPU.

The onnx model is attatched below:

Hi,

Could you also share the inference source code with us for checking?
Thanks.

Hi,
The code and model are attached below:

Hi,

Thanks for your sharing.
Will let you know for any progress.

Thanks.

OK, waiting for your reply.
Thanks.

Hi,

This issue can be reproduced in our environment.
We are passing this to our internal team for further suggestion.

Will share you more information once we got any feedback.

OK.
Thanks.

Hi,

This issue is fixed in our future release.
Will update here once it releases.

Thanks.

OK.
Hope to you release it soon.
Thanks.