Thanks for providing such powerful Tensorrt. In order to maximize the efficiency, we are using the dla with standalone mode, and using Int8 as input/output data type. Also we set the flag to allow all formats of input/output.
But we don’t know which type/format of input data we need to prepare.
1> When all of input is Int8, the input data in calibrator will be fp32 or fp16 or Int8?
2> Will trt reformat the input automatically?
Hi, Please refer to the below links to perform inference in INT8
@NVES Thanks for your reply.
By referencing the code A
and B, it looks like the input is always of FLOAT & LINEAR. But is this based on the premise of FLOAT & LINEAR network inputs or works on any input type & format?
Yes, It need to be float32. Please refer to the following doc for more details.
This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.