Hello NVIDIA Dev Team,
I’m an AI engineer and systems architect currently working on a next-generation AI assistant and local inference platform. This project is designed for real-time, on-device LLMs, computer vision, and multimodal processing—all running 100% locally on consumer-grade hardware with no reliance on external APIs or cloud backends.
My current build environment includes:
- GPU: NVIDIA RTX 5080 (Ada Gen3)
- CUDA Version: 12.8
- Linux: WSL2 + Ubuntu
- Build Target: PyTorch v2.5.0 from source
- Arch Flags:
sm_120
, which is currently unsupported by upstream select_compute_arch.cmake
After patching select_compute_arch.cmake
manually to support sm_120
, I’ve run into constant compatibility and build-breaking issues. It seems sm_120
is not yet integrated as a known architecture, despite this GPU family being commercially available for over a month.
My questions:
- When can we expect official support for
sm_120
in CUDA and PyTorch?
- Is there a preferred patch or integration pathway for developers building experimental CUDA/PyTorch environments for next-gen hardware?
- Can you confirm if
sm_120
is aligned with an upcoming compute capability update (e.g., 9.x, 10.x)?
I am happy to provide build logs, system diagnostics, and project goals if it helps expedite integration or beta access to upcoming CUDA revisions.
Thank you for your time and commitment to developers like me pushing the limits of local AI compute.
— The Architect
1 Like
The current nighly build of pytorch 2.8.0 works better with cuda 12.8. I dont think cuda 12.8 is supported at all in previous builds. Hopefully installing that will make some of your issues go away.
SUCCESS: PyTorch with CUDA 12.8 on RTX 5080 (sm_120) — Working
Following up with an important update:
I was able to successfully build PyTorch from source with full sm_120
support for my RTX 5080 using CUDA 12.8. This required a few custom modifications, but the result is stable and fully functional.
Summary of What Worked:
- PyTorch version: 2.5.0 (built from source)
- CUDA version: 12.8
- GPU architecture: sm_120 (manually patched)
- Platform: WSL2 + Ubuntu
- Compiler: gcc / g++ installed manually
- Build system: Ninja
Key Steps:
- Cloned PyTorch and checked out the v2.5.0 tag.
- Manually patched
select_compute_arch.cmake
to recognize sm_120
.
- Set
TORCH_CUDA_ARCH_LIST="12.0;12.1;12.2;12.3+PTX"
.
- Configured CMake with:
bash
CopyEdit
cmake .. -G Ninja \
-DCMAKE_BUILD_TYPE=Release \
-DUSE_CUDA=1 \
-DCUDA_TOOLKIT_ROOT_DIR=/usr/local/cuda \
-DTORCH_CUDA_ARCH_LIST="12.0;12.1;12.2;12.3+PTX" \
-DCMAKE_C_COMPILER=/usr/bin/gcc \
-DCMAKE_CXX_COMPILER=/usr/bin/g++ \
-DCMAKE_EXPORT_COMPILE_COMMANDS=ON
- Ran
python setup.py install
inside the PyTorch directory.
Result:
Confirmed success with:
python
CopyEdit
import torch
print(torch.cuda.get_device_properties(0))
Output:
scss
CopyEdit
_CudaDeviceProperties(name='NVIDIA GeForce RTX 5080', major=12, minor=0, total_memory=16302MB, multi_processor_count=84)
Let me know if anyone wants the patch or full build script. I’m happy to share.
— The Architect
2 Likes
Hi - yes please share the full build script. I’m hoping to use my 5080 to fine-tune using unsloth but running into issues. thanks in advance!