This doc state that orin NX support bfloat16
However it is not available when converting a ONNX model to tensorRT.
/usr/src/tensorrt/bin/trtexec --help
--noTF32 Disable tf32 precision (default is to enable tf32, in addition to fp32)
--fp16 Enable fp16 precision, in addition to fp32 (default = disabled)
--int8 Enable int8 precision, in addition to fp32 (default = disabled)
--best Enable all precisions to achieve the best performance (default = disabled)
Same here, only classic FP16 and no bfloat16.
https://docs.nvidia.com/deeplearning/tensorrt/api/python_api/infer/FoundationalTypes/DataType.html
How to enable TensoRT convewrsion in bfloat16 ?
