From the TensorRT document, there is no condition about using fp16. There is only condition for using int8 about gpu capability. I tried fp16 on the TITAN XP and 1080 Ti, but I failed in both. By the searching, I saw the statement: “You need volta for fp16”. This is right and I cannot use fp16 in my environment?
$ ./sample_int8 mnist
FP32 run:400 batches of size 100 starting at 100
........................................
Top1: 0.9904, Top5: 1
Processing 40000 images averaged 0.00181013 ms/image and 0.181013 ms/batch.
FP16 run:400 batches of size 100 starting at 100
Engine could not be created at this precision
INT8 run:400 batches of size 100 starting at 100
........................................
Top1: 0.9908, Top5: 1
Processing 40000 images averaged 0.00140439 ms/image and 0.140439 ms/batch.