there is no error message about dla after change above option dla=False. thanks.
but still process killed again and i think the reason of this problem is out of memory.
when i test on jetson nano 4GB, ( cli mode, available RAM size 3.2 GB ) operation work well.
but jetson orin nano 4GB, (cli mode, available RAM size 2.2GB) process killed.
thank you.
I tried the two methods you recommend, but swap memory didn’t solve this problem, probably for the reasons below.
but change the batch is worked. (change 8 to 4.) thanks.
but problem of change batch size is do inference 2 times.
it takes time spent 1.4 times more at inference.
i want keep my batch size for optimize my system
my question is,
Can you recommend another method for get more free RAM space?
(Stop certain services or other methods, etc. current available RAM size is 2.2GB in idle state, w/o GUI )
maybe 200MB space seems to be enough for pytorch model to TensorRT conversion and inference with keep my batchsize.
Loading cuDNN memory can take up to 600M or more.
There is a function called setTacticSources in TensorRT allows the user to deploy without calling cuDNN.
The function seems not been exported in torch2trt.
Is it possible to use TensorRT API directly?
This will require you to convert the model into ONNX format first.