Hi, I recently bought a Jetson Nano Development Kit and tried running local models for text generation on it. For example, Ollama works, but without CUDA support, it’s slower than on a Raspberry Pi! The Jetson Nano costs more than a typical Raspberry Pi, but without CUDA support, it feels like a total waste of money.
Is there a way to run these models with CUDA 10.2 support?
Unfortunately, we don’t have experience with llm on Jetson Nano with CUDA 10.2.
However, we have a generative AI tutorial for the Orin series (including Orin Nano).
If this is an option for you, please give it a try.