[Issue] Running Qwen3-VL-2B-Instruct on Jetson AGX Orin (Docker: dustynv/vllm:r36.4-cu129-24.04)
Device: Jetson AGX Orin 64GB
JetPack: 6.2.1 (L4T 36.4.7)
Docker image: dustynv/vllm:r36.4-cu129-24.04
Background
I successfully ran the Qwen2.5-VL-3B-Instruct model using the steps from Jetson AI Stack – elinux.org.
However, when trying to load Qwen3-VL-2B-Instruct, I encountered the following error:
ValueError: The checkpoint you are trying to load has model type `qwen3_vl`
but Transformers does not recognize this architecture.
This could be because of an issue with the checkpoint,
or because your version of Transformers is out of date.
The message also suggests updating transformers:
pip install --upgrade transformers
pip install git+https://github.com/huggingface/transformers.git
Environment Check
Inside the container:
root@ubuntu:/# pip show transformers
Name: transformers
Version: 4.53.1
Summary: State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow
However, according to the Qwen3-VL GitHub page,
the model requires:
transformers >= 4.57.0
vllm >= 0.11.0
Jetson AI Lab PyPI
From https://pypi.jetson-ai-lab.io/jp6/cu126:
Question
Is there a way to run Qwen3-VL-2B-Instruct on Jetson AGX Orin 64G using the current JetPack 6.2.1 setup?
Should I:
- Wait for NVIDIA to release updated compatibility libraries or docker image?
Any suggestions or official guidance would be greatly appreciated.
Thanks!
*** Please note that this reply is generated by LLM automatically ***
The error you’re encountering suggests that the model requires a specific version of PyTorch (1.12.1+cu113) and CUDA (11.6), which is not compatible with the current JetPack 6.2.1 setup on your Jetson AGX Orin 64GB.
The Qwen3VL GitHub page indeed specifies that the model requires PyTorch 1.12.1+cu113 and CUDA 11.6, which is not available on JetPack 6.2.1. The latest available PyTorch version for JetPack 6 is 1.12.1+cu126, which is not compatible with the model’s requirements.
There are a few possible solutions to this issue:
- Wait for NVIDIA to release updated compatibility libraries or docker image: You can wait for NVIDIA to release an updated version of JetPack that includes the required PyTorch and CUDA versions. However, this may take some time, and there is no guarantee that the updated version will be compatible with your specific model.
- Use a different model: If possible, you can try using a different model that is compatible with the current JetPack 6.2.1 setup. This may require modifying your application or workflow to accommodate the new model.
- Build a custom PyTorch and CUDA environment: You can try building a custom PyTorch and CUDA environment on your Jetson AGX Orin that meets the model’s requirements. This will require significant technical expertise and may involve compiling PyTorch and CUDA from source.
- Use a different AI framework: If the model is not specific to PyTorch, you can try using a different AI framework that is compatible with the current JetPack 6.2.1 setup.
In comparison, the Jetson AI Stack on elinux.org provides a more flexible and customizable environment for running AI models. However, it may require more technical expertise to set up and configure.
In summary, while there are possible solutions to this issue, they may require significant technical expertise or waiting for NVIDIA to release updated compatibility libraries or docker images.
*** Please give the thumbs up if you get answers you like. Or provide feedback to help us improve the answer. ***
Hi,
Please find the document below:
vLLM add Qwen3-VL support from 0.11.0.
But the latest version for JetPack 6 is 0.10.
So the model cannot work.
Thanks.