When I am using Nano, I want to run a VILA-7B model on it, but I can only use 6.3GB of RAM, so the process will be killed because of oom. How can I solve this problem. My code is :
python3 -m local_llm.chat --api=mlc --model /data/vila/vila_7b --vision-model /data/clip-vit-large-patch14-336 --max-context-len 1536 --max-new-tokens 128
Hi,
The app used 6.1G over a total of 6.3G memory, which is already 97%.
Please use the model that requires less memory.
Thanks.
But Nano should has 8GB Ram, why I can only see 6.3GB? How can I get more memory?
There is no update from you for a period, assuming this is not an issue any more.
Hence we are closing this topic. If need further support, please open a new one.
Thanks
Hi,
Could you also share the tegrastats log with us?
$ sudo tegrastats
Thanks.