JetPack Version : 4.4
The one-batch model running on the DLA in fp16 always comes out the wrong result, but the multi-batch infer result is different with the same input and only the first batch is wrong.
The model using one or more batches running normally on GPU.
The onnx model is attatched below:
https://drive.google.com/file/d/1jTuSd281-buxbf77sJb5YM3kdz8UQC6z/view?usp=sharing
Hi,
Could you also share the inference source code with us for checking?
Thanks.
Hi,
Thanks for your sharing.
Will let you know for any progress.
Thanks.
OK, waiting for your reply.
Thanks.
Hi,
This issue can be reproduced in our environment.
We are passing this to our internal team for further suggestion.
Will share you more information once we got any feedback.
Hi,
This issue is fixed in our future release.
Will update here once it releases.
Thanks.
OK.
Hope to you release it soon.
Thanks.
Hi,
Thanks for your patience.
This issue is fixed in our latest JetPack 4.6 release (TensorRT 8.0).
Please upgrade your environment to avoid this issue.
Thanks.
system
Closed
November 4, 2021, 3:08am
15
This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.