I am using the Jetson Orin Nano to run the live LLAVA video query with the nanodb integration as follows:
jetson-containers run $(autotag nano_llm)
python3 -m nano_llm.agents.video_query --api=mlc
–model Efficient-Large-Model/VILA1.5-3b
–max-context-len 256
–max-new-tokens 32
–video-input /dev/video0
–video-output webrtc://@:8554/output
–nanodb /data/nanodb/coco/2017
However, i notice that once i start to run the code, before the gstreamer gets connected, my orin nano will hang for long before crashing and rebooting by itself.
I have changed the settings to run it in MAXN super as well as ensuring that the RAM is only being used on the inference done from the code. however this does not seem to be enough for the orin nano, and i am unable to check the tegrastats or free -h stats as the orin nano freezes before i can even reach that page.
Any advice is appreciated. Thanks