Hi,
To inference a model with TensorRT, you will need to convert all the TensorFlow operations into TensorRT’s first.
Since the operations are not all one-to-one mapping, it needs some manually setting to wrap the operation.
As a result, please run the step.1 first.
It helps you to convert the operations into TensorRT API and serialize the output as a file.
After you got a TensorRT engine, you can deploy it with the sample shared in step.2.
It’s a python example to run a TensorRT engine with OpenCV image input.
And parse the output as the detection (bounding box+confidence).
Thanks.