for the VLM I have these volume settings, but I didn’t see this for the LLM
#### LOCAL MODEL STORE ON WORKER NODE
- name: local-model-store
hostPath:
path: /data/vila-1.5-40b:vila-yi-34b-siglip-stage3_1003_video_v8
#### LOCAL MODEL STORE IN POD
- name: local-model-store
mountPath: /tmp/vila-1.5-40b:vila-yi-34b-siglip-stage3_1003_video_v8
I have used the ngc registry command to download both the VLM and the LLM:
ngc registry model download-version "nim/nvidia/vila-1.5-40b:vila-yi-34b-siglip-stage3_1003_video_v8"
ngc registry model download-version "nvidia/nemo/llama-3_1-70b-instruct-nemo:1.0"
I just need to somehow modify the value.yaml file in the NIM LLM portion to be able to call the hostpath for the model I downloaded.
Correct, we need at least to have the large LLM and VLM locally saved for applications working in an air-gapped, on prem environment that cannot have access to public internet after deployment. Many of our customers will require this same feature, thank you for your help! :-)
Right but if we redeploy, it will require to download again. That is the problem. If we keep the PVC and don’t delete them, it won’t recognize the previous PVC, so in order to redeploy we have to delete all the previous PVC and start over again with a new download process.
If we can have some configuration for the LLM to host the model locally with hostPath or volumeMount - similar to the VLM, then this would solve the problem when we redeploy.
Is there a method to map a local model folder to the LLM in the VSS-Engine pod?