TensorFlow 2 Models for Jetson Nano


I’m currently working on an AI project using an NVIDIA Jetson Nano (4GB) and TensorFlow 2 where we were planning on using a Faster R-CNN Inception ResNet V2 640x640 model. We tried using TF-TRT to reduce the network, but it seems to be too big to fit in, the vRAM memory is not big enough and using Swap doesn’t solve the issue.

We have done several tests and, for the moment, the heaviest network from the TensorFlow Model Zoo we managed to get working is the SSD MobileNet V2 FPNLite 640x640.

I’ve been searching for a list of networks that have been tested on this device for TF2, but I can’t seem to find it. I know of the existence of this list, but it is for TF1 and doesn’t involve the TF2 Model Zoo models.

Is there any chance that a list of working models with their speed, memory usage and mAP tested on a common dataset (COCO or similar) will be developed? This would be specially interesting regarding the different methods available to deploy a model, either via CPU, optimized CPU with TFLite, GPU, optimized GPU with TF-TRT, fully optimized with pure TensorRT or any other possible option that I haven’t considered.

It seems to be a somewhat-common issue, all non-experts feel a bit lost, me included.

Thank you in advance.


It’s more recommended to use pure TensorRT.
Since TensorFlow keeps updating its operation, the layers that can be converted into TensorRT are extremely limited.

For TF2, you will need to convert the model into ONNX format first.
Below is the benchmark table of different models:
Please note that we don’t include the mAP score since it should be the same as CPU inference.

Here is an example to convert a TensorFlow model into TensorRT for your reference:


This topic was automatically closed 14 days after the last reply. New replies are no longer allowed.