Hi, I want to perform CPU inference using TAO-trained MaskRCNN models. I previously was able to do so using TAO-trained DetectNet_v2 models by converting them into ONNX, and I was hoping to do the same for TAO-trained MaskRCNN models.
I’m trying to do this as I want to use the ONNX weights with a CPU-only instance of Triton Inference Server. It currently works with the TAO-trained DetectNet_v2-converted ONNX models, and I’m trying to figure out how to do so for the TAO-trained MaskRCNN models as well.
I understand that back in 2022, there was no support for direct TAO-trained MaskRCNN → ONNX conversion. Does this still hold true as of today?
If yes, are there other, perhaps indirect ways to perform this conversion? (E.g. converting/exporting into an intermediary file format that can then be converted into ONNX?)
I get an UFF file instead of an ONNX file. Specifically, the logs show:
⋮
2024-10-02 04:12:25,307 [TAO Toolkit] [INFO] nvidia_tao_tf1.cv.mask_rcnn.export.exporter 300: Converted model was saved into /usr/src/app/mask_rcnn/experiment_dir_unpruned/model.epoch-199.uff
⋮
After checking, currently, Mask_rcnn does not support export to ONNX file yet. Suggest you to use trtexec to generate tensorrt engine for running inference on GPU.