What are the dim values representing in the .prototext files. For example:
/opt/nvidia/deepstream/deepstream/samples/models/Primary_Detector_Nano/resent10.prototext has dim values of 480x272.
and
/opt/nvidia/deepstream/deepstream/samples/models/Primary_Detector/resent10.prototext has dim values of 640x368
The performance section of the documentation suggests we lower these values to 480x272 on the jetson Xavier and NX. Why? does that make inference quicker if you specify the input dimensions to be smaller? At the expense of accuracy?
What would happen if we changed these values to higher one? Shouldn’t these dims match the size of the images the models were trained on?
Please provide complete information as applicable to your setup.
Hey Jason
The model is a fully convolutional network, it can support inferencing at image dimensions different than that of the dimensions it was trained on. So technically any input dimension greater than 16x16 (wxh) can work with the model.
Thanks @bcao - so why were the defaults of 480x272 and 640x368 chosen? For performance is it true that you make these smaller and in general if you make them bigger it will provide better accuracy?