I have a fresh install of the latest JetPack on a new Jetson Orin Nano Dev Kit Super. Installed it via the desktop flash method yesterday (02/27/2025).
When loading Ollama via Docker, the models, even smaller ones, load 100% CPU. I get the folloing messages in the Ollama logs:
time=2025-02-28T14:05:25.428Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths=[/usr/lib/aarch64-linux-gnu/nvidia/libcuda.so.1.1]
initializing /usr/lib/aarch64-linux-gnu/nvidia/libcuda.so.1.1
library /usr/lib/aarch64-linux-gnu/nvidia/libcuda.so.1.1 load err: /usr/lib/aarch64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by /usr/lib/aarch64-linux-gnu/nvidia/libnvrm_gpu.so)
time=2025-02-28T14:05:25.428Z level=INFO source=gpu.go:612 msg="Unable to load cudart library /usr/lib/aarch64-linux-gnu/nvidia/libcuda.so.1.1: Unable to load /usr/lib/aarch64-linux-gnu/nvidia/libcuda.so.1.1 library to query for Nvidia GPUs: /usr/lib/aarch64-linux-gnu/libc.so.6: version `GLIBC_2.34' not found (required by /usr/lib/aarch64-linux-gnu/nvidia/libnvrm_gpu.so)"
time=2025-02-28T14:05:25.430Z level=DEBUG source=gpu.go:574 msg="Unable to load cudart library /usr/lib/ollama/cuda_v11/libcudart.so.11.3.109: your nvidia driver is too old or missing. If you have a CUDA GPU please upgrade to run ollama"
My docker-compose.yaml is farly simple, following the basic example:
According to ldd, I am running GLIBC_2.35, so I would assume I meet the requirements.
ldd (Ubuntu GLIBC 2.35-0ubuntu3.8) 2.35
Copyright (C) 2022 Free Software Foundation, Inc.
This is free software; see the source for copying conditions. There is NO
warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
Written by Roland McGrath and Ulrich Drepper.
As far as I know, everything is current, aside from Docker which I am purposely holding back to 27.x because of issues with 28.x.
I’ve messed with this all weekend, with no success.
When I try to run Ollama with jetson-containers, I still have an issue with the CUDA libraries, no GPU is found, and the Ollama models run in 100% CPU.
I ran Ollama 0.5.12 (latest, at the time) and worked backwards through the versions until I found 0.5.7 worked as expected. Everything 0.5.8 and newer reports:
ollama | time=2025-03-04T20:48:09.206Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths="[/usr/lib/ollama/cuda_v11/libcudart.so.11.3.109 /usr/lib/ollama/cuda_v12/libcudart.so.12.4.127]"
ollama | cudaSetDevice err: 35
ollama | time=2025-03-04T20:48:09.209Z level=DEBUG source=gpu.go:574 msg="Unable to load cudart library /usr/lib/ollama/cuda_v11/libcudart.so.11.3.109: your nvidia driver is too old or missing. If you have a CUDA GPU please upgrade to run ollama"
ollama | cudaSetDevice err: 35
ollama | time=2025-03-04T20:48:09.210Z level=DEBUG source=gpu.go:574 msg="Unable to load cudart library /usr/lib/ollama/cuda_v12/libcudart.so.12.4.127: your nvidia driver is too old or missing. If you have a CUDA GPU please upgrade to run ollama"
ollama | time=2025-03-04T20:48:09.210Z level=DEBUG source=amd_linux.go:419 msg="amdgpu driver not detected /sys/module/amdgpu"
ollama | time=2025-03-04T20:48:09.210Z level=INFO source=gpu.go:377 msg="no compatible GPUs were discovered"
ollama | time=2025-03-04T20:48:09.211Z level=INFO source=types.go:130 msg="inference compute" id=0 library=cpu variant="" compute="" driver=0.0 name="" total="7.4 GiB" available="6.5 GiB"