Hi, I’m doing facial expression recognition but I want to inference the model on my jetson nano, so I converted keras model (.h5) to tensorflow frozen model (.pb) then optimized it with tensorrt , finally I got trt_graph.pb.
this is the code for converting .h5 to .pb
this is the .pb model I generated using above script.
I tried to use this block of code but it’s extremely slow, it takes more than 15-20s just to predict a single image.
My question is : I want to use tensorrt to inference the .pb model on my jetson nano but I don’t know how to do it. :( ,
Any guidelines or samples code would be appreciate.
TensorRT Version : 188.8.131.52-1
Operating System + Version : ubuntu 20.04
Python Version (if applicable) : 3.6
TensorFlow Version (if applicable) : 1.15