Invalid argument: Input shapes are inconsistent on the batch dimension, for TRTEngineOp_0(for secondary inference))

Hi @bcao,
I got my model optimized offline and it worked in deepstream pipeline. Although the original model with changes to the config.pbtxt still does not work.

The following post(section about optimization) was helpful for the conversion of FP32 TF model to optimized FP16 TF-TRT graph:
https://developer.nvidia.com/blog/deploying-models-from-tensorflow-model-zoo-using-deepstream-and-triton-inference-server/