Good day, hopefully this is the proper location for this issue. I’ve found other threads with a similar problem but they’re either dated or I don’t understand the solution that was given.
I’m attempting to install xformers on Nobara linux (Fedora-based) and I’m getting this error as I attempt to install it:
pip install -e .
Obtaining file:///home/kymus/stable-diffusion-webui/repositories/xformers
Preparing metadata (setup.py) ... done
Requirement already satisfied: torch>=1.12 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (1.13.1+cu117)
Requirement already satisfied: numpy in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (1.23.3)
Requirement already satisfied: pyre-extensions==0.0.23 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (0.0.23)
Requirement already satisfied: typing-inspect in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (0.8.0)
Requirement already satisfied: typing-extensions in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (4.4.0)
Requirement already satisfied: mypy-extensions>=0.3.0 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from typing-inspect->pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (1.0.0)
Installing collected packages: xformers
Running setup.py develop for xformers
error: subprocess-exited-with-error
× python setup.py develop did not run successfully.
│ exit code: 1
╰─> [76 lines of output]
No ROCm runtime is found, using ROCM_HOME='/opt/rocm-5.2.3'
running develop
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/easy_install.py:156: EasyInstallDeprecationWarning: easy_install command is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
running egg_info
writing xformers.egg-info/PKG-INFO
writing dependency_links to xformers.egg-info/dependency_links.txt
writing requirements to xformers.egg-info/requires.txt
writing top-level names to xformers.egg-info/top_level.txt
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:476: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend.
warnings.warn(msg.format('we could not find ninja.'))
reading manifest file 'xformers.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file 'xformers.egg-info/SOURCES.txt'
running build_ext
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:387: UserWarning: The detected CUDA version (11.8) has a minor version mismatch with the version that was used to compile PyTorch (11.7). Most likely this shouldn't be a problem.
warnings.warn(CUDA_MISMATCH_WARN.format(cuda_str_version, torch.version.cuda))
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:397: UserWarning: There are no g++ version bounds defined for CUDA version 11.8
warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}')
building 'xformers._C_flashattention' extension
gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -fPIC -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/fmha_api.cpp -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/fmha_api.o -O3 -fopenmp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
In file included from /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:42,
from third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:33:
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h: In function ‘void set_alpha(uint32_t&, float, Data_type)’:
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:63:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
63 | alpha = reinterpret_cast<const uint32_t &>( h2 );
| ^~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:68:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
68 | alpha = reinterpret_cast<const uint32_t &>( h2 );
| ^~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:70:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
70 | alpha = reinterpret_cast<const uint32_t &>( norm );
| ^~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘void set_params_fprop(FMHA_fprop_params&, size_t, size_t, size_t, size_t, size_t, at::Tensor, at::Tensor, at::Tensor, at::Tensor, void*, void*, void*, void*, void*, float, float, bool, int)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:64:11: warning: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type ‘struct FMHA_fprop_params’; use assignment or value-initialization instead [-Wclass-memaccess]
64 | memset(¶ms, 0, sizeof(params));
| ~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:75:8: note: ‘struct FMHA_fprop_params’ declared here
75 | struct FMHA_fprop_params : public Qkv_params {
| ^~~~~~~~~~~~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:60:15: warning: unused variable ‘acc_type’ [-Wunused-variable]
60 | Data_type acc_type = DATA_TYPE_FP32;
| ^~~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_fwd(const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, bool, bool, int, c10::optional<at::Generator>)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:208:10: warning: unused variable ‘is_sm80’ [-Wunused-variable]
208 | bool is_sm80 = dprops->major == 8 && dprops->minor == 0;
| ^~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_bwd_block(const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, c10::optional<at::Generator>)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:652:10: warning: unused variable ‘is_sm8x’ [-Wunused-variable]
652 | bool is_sm8x = dprops->major == 8 && dprops->minor >= 0;
| ^~~~~~~
/usr/local/cuda/bin/nvcc -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.cu -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options '-fPIC' -DHAS_PYTORCH --use_fast_math -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ --extended-lambda -D_ENABLE_EXTENDED_ALIGNED_STORAGE --generate-line-info -DNDEBUG --threads 4 --ptxas-options=-v -O3 -std=c++17 --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -gencode=arch=compute_75,code=sm_75 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 --generate-line-info -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
error: command '/usr/local/cuda/bin/nvcc' failed with exit code 255
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.
error: subprocess-exited-with-error
× python setup.py develop did not run successfully.
│ exit code: 1
╰─> [76 lines of output]
No ROCm runtime is found, using ROCM_HOME='/opt/rocm-5.2.3'
running develop
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/easy_install.py:156: EasyInstallDeprecationWarning: easy_install command is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools.
warnings.warn(
running egg_info
writing xformers.egg-info/PKG-INFO
writing dependency_links to xformers.egg-info/dependency_links.txt
writing requirements to xformers.egg-info/requires.txt
writing top-level names to xformers.egg-info/top_level.txt
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:476: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend.
warnings.warn(msg.format('we could not find ninja.'))
reading manifest file 'xformers.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
adding license file 'LICENSE'
writing manifest file 'xformers.egg-info/SOURCES.txt'
running build_ext
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:387: UserWarning: The detected CUDA version (11.8) has a minor version mismatch with the version that was used to compile PyTorch (11.7). Most likely this shouldn't be a problem.
warnings.warn(CUDA_MISMATCH_WARN.format(cuda_str_version, torch.version.cuda))
/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:397: UserWarning: There are no g++ version bounds defined for CUDA version 11.8
warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}')
building 'xformers._C_flashattention' extension
gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -fPIC -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/fmha_api.cpp -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/fmha_api.o -O3 -fopenmp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
In file included from /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:42,
from third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:33:
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h: In function ‘void set_alpha(uint32_t&, float, Data_type)’:
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:63:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
63 | alpha = reinterpret_cast<const uint32_t &>( h2 );
| ^~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:68:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
68 | alpha = reinterpret_cast<const uint32_t &>( h2 );
| ^~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:70:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
70 | alpha = reinterpret_cast<const uint32_t &>( norm );
| ^~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘void set_params_fprop(FMHA_fprop_params&, size_t, size_t, size_t, size_t, size_t, at::Tensor, at::Tensor, at::Tensor, at::Tensor, void*, void*, void*, void*, void*, float, float, bool, int)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:64:11: warning: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type ‘struct FMHA_fprop_params’; use assignment or value-initialization instead [-Wclass-memaccess]
64 | memset(¶ms, 0, sizeof(params));
| ~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~
/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:75:8: note: ‘struct FMHA_fprop_params’ declared here
75 | struct FMHA_fprop_params : public Qkv_params {
| ^~~~~~~~~~~~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:60:15: warning: unused variable ‘acc_type’ [-Wunused-variable]
60 | Data_type acc_type = DATA_TYPE_FP32;
| ^~~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_fwd(const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, bool, bool, int, c10::optional<at::Generator>)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:208:10: warning: unused variable ‘is_sm80’ [-Wunused-variable]
208 | bool is_sm80 = dprops->major == 8 && dprops->minor == 0;
| ^~~~~~~
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_bwd_block(const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, c10::optional<at::Generator>)’:
third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:652:10: warning: unused variable ‘is_sm8x’ [-Wunused-variable]
652 | bool is_sm8x = dprops->major == 8 && dprops->minor >= 0;
| ^~~~~~~
/usr/local/cuda/bin/nvcc -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.cu -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options '-fPIC' -DHAS_PYTORCH --use_fast_math -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ --extended-lambda -D_ENABLE_EXTENDED_ALIGNED_STORAGE --generate-line-info -DNDEBUG --threads 4 --ptxas-options=-v -O3 -std=c++17 --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -gencode=arch=compute_75,code=sm_75 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 --generate-line-info -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
In file included from /usr/local/cuda/include/cuda_runtime.h:83,
from <command-line>:
/usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
| ^~~~~
error: command '/usr/local/cuda/bin/nvcc' failed with exit code 255
[end of output]
note: This error originates from a subprocess, and is likely not a problem with pip.