I want to use TF-TRT Python API to optimize the graph,
and then use TF C++ API for deployment on NVIDIA Xavier.
Is TF C++ API capable of running TF-TRT optimized graph?
Also, what is the preferred way of deploying TRT optimized model on Jetson?
According to item 8.1 of https://docs.nvidia.com/deeplearning/sdk/tensorrt-developer-guide/index.html
“Note: The UFF Parser which is used to parse a network in UFF format will be deprecated in a future release. The recommended method of importing TensorFlow models to TensorRT is using TensorFlow with TensorRT (TF-TRT).”
While item 7.2 says to use TRT C++ API with UFF as an intermediate format.