What is the proposed workflow for converting a model (SavedModel,…) that was trained under Linux to a TensorRT engine that can be deployed under Windows?
As far as i understand the TensorFlow with TensorRT (TF-TRT) is not supported under Windows. Running this optimization scheme under Linux and trying to use the serialized engine / plan under Windows ist not possible as engines/plans are not portable across OSs or different hardware. See support matrix.
So I guess the only possibility is to use an intermediate format (UFF,ONNX) to do the conversion under Linux and then use the TRT UFF/ONNX parsers under Windows to generate the engine? However, this seems like a rather cumbersome approach that may also suffer from incompatibilities or missing features between the different model representations. Are there alternative workflows or strategies?