I tried to run my Tensorflow-Keras-Unet Model on Jetson-Nano and Jetson-TX2 ,
but the inference time was so worse in the range of 1.5 minutes for a single image inference.
the model was trained on a windows machine and works fine in windows and ubuntu18 with an inference time in the range of 10us to 19ms
I tried running the Jetson Boards (Nano&TX2) at Max Performance modes but no response of improvement . Still inference time as 90s to 100s
I tried with TensorRT Optimization Techniques but I couldn’t complete the process as lack of knowledge about the same I couldn’t found a resource to learn about this optimization for the Image Semantic Segmentation Problems (say Unet)
Can You Provide some resources to know about this more or any other methods to improve the inference latency