I am using a Jetson Orin AGX 64GB with the following environment:
Device: Jetson Orin AGX 64GB
JetPack: 6.2 (L4T r36.4.3)
CUDA: 12.6
OS: Ubuntu 22.04 (aarch64)
I am trying to run the DeepSeek-OCR project on this device, which relies on vLLM.
However, I am having difficulty finding a compatible vLLM version for this platform.
*** Please note that this reply is generated by LLM automatically ***
The question is about building a specific version of vLLM (version 0.11.1) from source on a Jetson Orin AGX 64GB device with CUDA 12.6. The user is experiencing difficulties finding a compatible prebuilt vLLM version for their platform and is seeking guidance on building vLLM from source.
To address this question, we need to consider the following points:
vLLM version compatibility: The user is trying to run the DeepSeekOCR project, which relies on vLLM 0.8.5. However, this version is not compatible with the Jetson Orin AGX 64GB device. The user has tried newer versions of vLLM, such as 0.11.x, but they are not available as prebuilt wheels for JetPack 6.2 and CUDA 12.6.
Building vLLM from source: The user is asking whether it is possible to build vLLM 0.11.1 from source on the Jetson Orin AGX 64GB device. This requires considering the build dependencies, patches, and CMake options required for aarch64 architecture.
Official support and recommendations: The user is seeking guidance on whether NVIDIA plans to provide additional prebuilt vLLM wheels or images for intermediate versions like 0.11.x.
To answer these questions, we can provide the following information:
Building vLLM from source: Yes, it is possible to build vLLM from source on the Jetson Orin AGX 64GB device. However, this requires careful consideration of the build dependencies, patches, and CMake options required for aarch64 architecture. The user can refer to the vLLM GitHub repository for build instructions and dependencies.
vLLM version compatibility: The user can try building vLLM 0.11.1 from source, but they should be aware that this version may not be compatible with the DeepSeekOCR project. The user may need to modify the DeepSeekOCR code to work with the newer vLLM version.
Official support and recommendations: NVIDIA does provide prebuilt vLLM wheels and images for various platforms, including JetPack 6.2 and CUDA 12.6. However, the availability of these prebuilt versions may depend on the specific vLLM version and platform. The user can check the NVIDIA website for the latest prebuilt vLLM versions and images.
In summary, the user can try building vLLM 0.11.1 from source on the Jetson Orin AGX 64GB device, but they should be aware of the potential compatibility issues with the DeepSeekOCR project. The user can also check the NVIDIA website for prebuilt vLLM versions and images for their platform.
Comparison with other options:
vLLM prebuilt wheels: The user has tried using prebuilt vLLM wheels from PyPI, but these wheels are not compatible with the Jetson Orin AGX 64GB device.
vLLM container images: The user has tried using official vLLM container images, but these images are not compatible with the DeepSeekOCR project.
Building vLLM from source: Building vLLM from source is a viable option, but it requires careful consideration of the build dependencies, patches, and CMake options required for aarch64 architecture.
Recommendations:
The user should try building vLLM 0.11.1 from source on the Jetson Orin AGX 64GB device, following the build instructions and dependencies provided in the vLLM GitHub repository.
The user should be aware of the potential compatibility issues with the DeepSeekOCR project and may need to modify the code to work with the newer vLLM version.
The user can check the NVIDIA website for prebuilt vLLM versions and images for their platform, but the availability of these prebuilt versions may depend on the specific vLLM version and platform.
*** Please give the thumbs up if you get answers you like. Or provide feedback to help us improve the answer. ***
Iβve tried this way for a moment.I used jetson-container and use the image tag
0.8.6-r36.4-cu128-24.04
and tried DeepSeek-OCR.It threw out an error:DeepseekOCRMultiModalProcessor._cached_apply_hf_processor()
got an unexpected keyword argument βreturn_mm_hashesβ.
Then I used the command to look up which version I can build.
cd jetson-containers
./build.sh vllm --list
I found that only 0.14.0 can be built.Iβve tried this whl package and failed before.
Could you check the minimum vLLM required by your model and build it from the source directly?
The repo shared above should be able to build different vLLM versions for Orin.
The minimum requirement for Deepseek-OCR may be cuda11.8+vllm0.8.5(This is the experimental environment that the project team used).Iβve tried the whl packages downloaded from jp6/cu126 index but it didnβt work for me.Next I will try to use the Dockerfile in jetson-container to build the version of vllm:v0.8.5.post1.Thanks.