Hello
I’m trying to use the NVILA-8B model from the Efficient-Large-Model repository(Efficient-Large-Model/NVILA-8B · Hugging Face) on a Jetson AGX Orin with NanoLLM. However, when I run the following command, the model fails to start:
jetson-containers run $(autotag nano_llm) \
python3 -m nano_llm.chat --api=mlc \
--model Efficient-Large-Model/NVILA-8b
I suspect there isn’t a Docker image that currently includes support for NVILA-8B out of the box.
Does anyone know if there is a prebuilt Docker image that can run NVILA-8B on Jetson AGX Orin, or how to build/configure one so that NVILA-8B can be used with the MLC back end? Any help or instructions would be greatly appreciated.
Thank you!