I am using this command to profile engine
/usr/src/tensorrt/bin/trtexec --loadEngine=<engine> --exportLayerInfo=graph.json --exportProfile=profile.json --warmUp=0 --duration=0 --iterations=10
And I got inference time from profile.json
, but I dont know this inference time inlcudes engine loading time or not. Please let me know. Thanks.