Hello,
Recently,I used the tensorrt2 to speed up a network,and compared its inference time with the origin network.I found that when testing on Titan X the engine doubles the inference speed.While on Jetson TX2,the engine can reach nearly 10 speedup ratio compared with the origin network.On Titan X,the platform doesn't support 16-bit khalf mode,and I used this mode on TX2.Is this the only reason that makes the engine can reach much bigger speedup ratio on TX2? Is there any other reasons? Waiting for your reply,thanks.