Pretrained Obect detection in TAO Yolo4 -> RT-DETR -> Dino?

TLDR: Should I train a Yolo4, RT-DETR, or DINO for object detection, given that I don’t have a powerful GPU?

I needed to make a new model, since the old Caffemodel we ran on the OG Jetson Nano does not work on my Jetson Xavier, as it uses a newer version of Deepstream.

I have about 5000 training images in KITTI format, and only one class of object to detect.

I was able to install TAO CLI and install the TAO tutorials. So first, I started the setup for training a Yolo4 model, but I didn’t notice it was deprecated until I couldn’t find the backbone to download from NGC. (I now know that the backbones are here)

From what I understand, based on the migration documentation for TOA 5.5 to 6.0, YOLO and a whole host of other models have been deprecated in favor of RT-DETR.

Now, there wasn’t any tutorial for RT-DETR in the TAO tutorials repo that I could find, but I imagine I should be able to get it done once I figure out where the pretrained_model / backbones are. I didn’t see anything that I thought might be them when using ngc registry model list. But there was documentation for training RT-DETR here.

There are Tutorials for DINO in the repo, so I assume it would not be a problem to get it to work. My only concern is that I don’t have a powerful enough GPU available, so I would be sitting around for a couple of weeks while it trains.

So, should I ignore that Yolo is deprecated and see if I can get it to work by downloading the backbone manually without NGC CLI? Should I try using RT-DETR, and if so, does anyone know where to download the backbone / pretrained model? Is using DINO feasible, or should I not try, given the hardware I have available?

From what I understand, all the models should be able to run on DeepStream. Thought, do correct me if I am wrong?

• Dev: Nvidia RTX 2080 Super / TAO 6.25

• Deploy: Hardware Jetson Xavier / Deepstream 6.3

• Network Type Yolo_v4 / RT-DETR / DINO

Although it is deprecated in latest TAO, you can still train it with an older version of TAO docker. You can find all TAO dockers in https://catalog.ngc.nvidia.com/orgs/nvidia/teams/tao/containers/tao-toolkit/tags. You can select nvcr.io/nvidia/tao/tao-toolkit:5.0.0-tf1.15.5 to run YOLOv4.

Yes, there is document for using RT-DETR. You can also search/find some users’ spec files in this TAO forum.

You can find info about pretrained models or supported backbones in Unable to reproduce RT-DETR original results on COCO.

The 8GB GPU memory is hard to train Dino. But you can have a try. If meet an OOM issue, please modify the input-size and retry. Strongly suggest you to use a gpu with more GPU memory.

Thanks for the help.

Got Yolo_v4 training stated by manually downloading the backbone from this catalog. Then adding the path, and arch to the spec files.

TAO CLI automatically downloaded the tao-toolkit:5.0.0-tf1.15.5 when I used the tao model yolo_v4 command.

1 Like