Hello, I am trying to get MobileNet V1 inference results in fp16 on the Jetson Nano following the instructions here: https://github.com/NVIDIA-AI-IOT/tf_trt_models
It seems like Mobilenet-V1 224-224 input is a valid download option, and indeed I downloaded it and the classification.py script runs fine. However, on the Jetson Nano, it gives an out of memory error, and indicates that the TensorRT optimization has failed. The batch 1 inference time is 0.15 seconds, which equates to 7FPS, suggesting quite poor performance.
Did anybody run into similar problems? What should I expect to be the FPS for batch 1 fp16 inference for Mobilenet V1 with 224-224 input? I am just running image classification, not detection.
Thank you for help!