Unsupported GNU version error

Good day, hopefully this is the proper location for this issue. I’ve found other threads with a similar problem but they’re either dated or I don’t understand the solution that was given.

I’m attempting to install xformers on Nobara linux (Fedora-based) and I’m getting this error as I attempt to install it:

pip install -e .
Obtaining file:///home/kymus/stable-diffusion-webui/repositories/xformers
  Preparing metadata (setup.py) ... done
Requirement already satisfied: torch>=1.12 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (1.13.1+cu117)
Requirement already satisfied: numpy in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (1.23.3)
Requirement already satisfied: pyre-extensions==0.0.23 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from xformers==0.0.17+836ab28.d20230222) (0.0.23)
Requirement already satisfied: typing-inspect in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (0.8.0)
Requirement already satisfied: typing-extensions in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (4.4.0)
Requirement already satisfied: mypy-extensions>=0.3.0 in /home/kymus/stable-diffusion-webui/venv/lib/python3.10/site-packages (from typing-inspect->pyre-extensions==0.0.23->xformers==0.0.17+836ab28.d20230222) (1.0.0)
Installing collected packages: xformers
  Running setup.py develop for xformers
    error: subprocess-exited-with-error
    
    × python setup.py develop did not run successfully.
    │ exit code: 1
    ╰─> [76 lines of output]
        No ROCm runtime is found, using ROCM_HOME='/opt/rocm-5.2.3'
        running develop
        /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/easy_install.py:156: EasyInstallDeprecationWarning: easy_install command is deprecated. Use build and pip and other standards-based tools.
          warnings.warn(
        /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools.
          warnings.warn(
        running egg_info
        writing xformers.egg-info/PKG-INFO
        writing dependency_links to xformers.egg-info/dependency_links.txt
        writing requirements to xformers.egg-info/requires.txt
        writing top-level names to xformers.egg-info/top_level.txt
        /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:476: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend.
          warnings.warn(msg.format('we could not find ninja.'))
        reading manifest file 'xformers.egg-info/SOURCES.txt'
        reading manifest template 'MANIFEST.in'
        adding license file 'LICENSE'
        writing manifest file 'xformers.egg-info/SOURCES.txt'
        running build_ext
        /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:387: UserWarning: The detected CUDA version (11.8) has a minor version mismatch with the version that was used to compile PyTorch (11.7). Most likely this shouldn't be a problem.
          warnings.warn(CUDA_MISMATCH_WARN.format(cuda_str_version, torch.version.cuda))
        /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:397: UserWarning: There are no g++ version bounds defined for CUDA version 11.8
          warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}')
        building 'xformers._C_flashattention' extension
        gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -fPIC -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/fmha_api.cpp -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/fmha_api.o -O3 -fopenmp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
        In file included from /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:42,
                         from third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:33:
        /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h: In function ‘void set_alpha(uint32_t&, float, Data_type)’:
        /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:63:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
           63 |         alpha = reinterpret_cast<const uint32_t &>( h2 );
              |                                                     ^~
        /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:68:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
           68 |         alpha = reinterpret_cast<const uint32_t &>( h2 );
              |                                                     ^~
        /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:70:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
           70 |         alpha = reinterpret_cast<const uint32_t &>( norm );
              |                                                     ^~~~
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘void set_params_fprop(FMHA_fprop_params&, size_t, size_t, size_t, size_t, size_t, at::Tensor, at::Tensor, at::Tensor, at::Tensor, void*, void*, void*, void*, void*, float, float, bool, int)’:
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:64:11: warning: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type ‘struct FMHA_fprop_params’; use assignment or value-initialization instead [-Wclass-memaccess]
           64 |     memset(&params, 0, sizeof(params));
              |     ~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~
        /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:75:8: note: ‘struct FMHA_fprop_params’ declared here
           75 | struct FMHA_fprop_params : public Qkv_params {
              |        ^~~~~~~~~~~~~~~~~
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:60:15: warning: unused variable ‘acc_type’ [-Wunused-variable]
           60 |     Data_type acc_type = DATA_TYPE_FP32;
              |               ^~~~~~~~
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_fwd(const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, bool, bool, int, c10::optional<at::Generator>)’:
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:208:10: warning: unused variable ‘is_sm80’ [-Wunused-variable]
          208 |     bool is_sm80 = dprops->major == 8 && dprops->minor == 0;
              |          ^~~~~~~
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_bwd_block(const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, c10::optional<at::Generator>)’:
        third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:652:10: warning: unused variable ‘is_sm8x’ [-Wunused-variable]
          652 |     bool is_sm8x = dprops->major == 8 && dprops->minor >= 0;
              |          ^~~~~~~
        /usr/local/cuda/bin/nvcc -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.cu -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options '-fPIC' -DHAS_PYTORCH --use_fast_math -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ --extended-lambda -D_ENABLE_EXTENDED_ALIGNED_STORAGE --generate-line-info -DNDEBUG --threads 4 --ptxas-options=-v -O3 -std=c++17 --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -gencode=arch=compute_75,code=sm_75 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 --generate-line-info -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0
        In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                         from <command-line>:
        /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
              |  ^~~~~
        In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                         from <command-line>:
        /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
              |  ^~~~~
        In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                         from <command-line>:
        /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
              |  ^~~~~
        In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                         from <command-line>:
        /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
              |  ^~~~~
        error: command '/usr/local/cuda/bin/nvcc' failed with exit code 255
        [end of output]
    
    note: This error originates from a subprocess, and is likely not a problem with pip.
error: subprocess-exited-with-error

× python setup.py develop did not run successfully.
│ exit code: 1
╰─> [76 lines of output]
    No ROCm runtime is found, using ROCM_HOME='/opt/rocm-5.2.3'
    running develop
    /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/easy_install.py:156: EasyInstallDeprecationWarning: easy_install command is deprecated. Use build and pip and other standards-based tools.
      warnings.warn(
    /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/setuptools/command/install.py:34: SetuptoolsDeprecationWarning: setup.py install is deprecated. Use build and pip and other standards-based tools.
      warnings.warn(
    running egg_info
    writing xformers.egg-info/PKG-INFO
    writing dependency_links to xformers.egg-info/dependency_links.txt
    writing requirements to xformers.egg-info/requires.txt
    writing top-level names to xformers.egg-info/top_level.txt
    /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:476: UserWarning: Attempted to use ninja as the BuildExtension backend but we could not find ninja.. Falling back to using the slow distutils backend.
      warnings.warn(msg.format('we could not find ninja.'))
    reading manifest file 'xformers.egg-info/SOURCES.txt'
    reading manifest template 'MANIFEST.in'
    adding license file 'LICENSE'
    writing manifest file 'xformers.egg-info/SOURCES.txt'
    running build_ext
    /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:387: UserWarning: The detected CUDA version (11.8) has a minor version mismatch with the version that was used to compile PyTorch (11.7). Most likely this shouldn't be a problem.
      warnings.warn(CUDA_MISMATCH_WARN.format(cuda_str_version, torch.version.cuda))
    /home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/utils/cpp_extension.py:397: UserWarning: There are no g++ version bounds defined for CUDA version 11.8
      warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}')
    building 'xformers._C_flashattention' extension
    gcc -Wno-unused-result -Wsign-compare -DDYNAMIC_ANNOTATIONS_ENABLED=1 -DNDEBUG -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -O2 -fexceptions -g -grecord-gcc-switches -pipe -Wall -Werror=format-security -Wp,-D_FORTIFY_SOURCE=2 -Wp,-D_GLIBCXX_ASSERTIONS -fstack-protector-strong -m64 -mtune=generic -fasynchronous-unwind-tables -fstack-clash-protection -fcf-protection -D_GNU_SOURCE -fPIC -fwrapv -fPIC -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/fmha_api.cpp -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/fmha_api.o -O3 -fopenmp -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0 -std=c++14
    In file included from /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:42,
                     from third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:33:
    /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h: In function ‘void set_alpha(uint32_t&, float, Data_type)’:
    /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:63:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
       63 |         alpha = reinterpret_cast<const uint32_t &>( h2 );
          |                                                     ^~
    /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:68:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
       68 |         alpha = reinterpret_cast<const uint32_t &>( h2 );
          |                                                     ^~
    /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha_utils.h:70:53: warning: dereferencing type-punned pointer will break strict-aliasing rules [-Wstrict-aliasing]
       70 |         alpha = reinterpret_cast<const uint32_t &>( norm );
          |                                                     ^~~~
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘void set_params_fprop(FMHA_fprop_params&, size_t, size_t, size_t, size_t, size_t, at::Tensor, at::Tensor, at::Tensor, at::Tensor, void*, void*, void*, void*, void*, float, float, bool, int)’:
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:64:11: warning: ‘void* memset(void*, int, size_t)’ clearing an object of non-trivial type ‘struct FMHA_fprop_params’; use assignment or value-initialization instead [-Wclass-memaccess]
       64 |     memset(&params, 0, sizeof(params));
          |     ~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~
    /home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src/fmha.h:75:8: note: ‘struct FMHA_fprop_params’ declared here
       75 | struct FMHA_fprop_params : public Qkv_params {
          |        ^~~~~~~~~~~~~~~~~
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:60:15: warning: unused variable ‘acc_type’ [-Wunused-variable]
       60 |     Data_type acc_type = DATA_TYPE_FP32;
          |               ^~~~~~~~
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_fwd(const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, bool, bool, int, c10::optional<at::Generator>)’:
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:208:10: warning: unused variable ‘is_sm80’ [-Wunused-variable]
      208 |     bool is_sm80 = dprops->major == 8 && dprops->minor == 0;
          |          ^~~~~~~
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp: In function ‘std::vector<at::Tensor> mha_bwd_block(const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, at::Tensor&, at::Tensor&, at::Tensor&, const at::Tensor&, const at::Tensor&, const at::Tensor&, int, int, float, float, bool, c10::optional<at::Generator>)’:
    third_party/flash-attention/csrc/flash_attn/fmha_api.cpp:652:10: warning: unused variable ‘is_sm8x’ [-Wunused-variable]
      652 |     bool is_sm8x = dprops->major == 8 && dprops->minor >= 0;
          |          ^~~~~~~
    /usr/local/cuda/bin/nvcc -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/flash-attention/csrc/flash_attn/src -I/home/kymus/stable-diffusion-webui/repositories/xformers/third_party/cutlass/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/TH -I/home/kymus/stable-diffusion-webui/venv/lib64/python3.10/site-packages/torch/include/THC -I/usr/local/cuda/include -I/home/kymus/stable-diffusion-webui/venv/include -I/usr/include/python3.10 -c third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.cu -o build/temp.linux-x86_64-3.10/third_party/flash-attention/csrc/flash_attn/src/fmha_block_dgrad_fp16_kernel_loop.sm80.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options '-fPIC' -DHAS_PYTORCH --use_fast_math -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ --extended-lambda -D_ENABLE_EXTENDED_ALIGNED_STORAGE --generate-line-info -DNDEBUG --threads 4 --ptxas-options=-v -O3 -std=c++17 --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -gencode=arch=compute_75,code=sm_75 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 --generate-line-info -DTORCH_API_INCLUDE_EXTENSION_H -DPYBIND11_COMPILER_TYPE="_gcc" -DPYBIND11_STDLIB="_libstdcpp" -DPYBIND11_BUILD_ABI="_cxxabi1011" -DTORCH_EXTENSION_NAME=_C_flashattention -D_GLIBCXX_USE_CXX11_ABI=0
    In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                     from <command-line>:
    /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
      132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          |  ^~~~~
    In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                     from <command-line>:
    /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
      132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          |  ^~~~~
    In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                     from <command-line>:
    /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
      132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          |  ^~~~~
    In file included from /usr/local/cuda/include/cuda_runtime.h:83,
                     from <command-line>:
    /usr/local/cuda/include/crt/host_config.h:132:2: error: #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
      132 | #error -- unsupported GNU version! gcc versions later than 11 are not supported! The nvcc flag '-allow-unsupported-compiler' can be used to override this version check; however, using an unsupported host compiler may cause compilation failure or incorrect run time execution. Use at your own risk.
          |  ^~~~~
    error: command '/usr/local/cuda/bin/nvcc' failed with exit code 255
    [end of output]

note: This error originates from a subprocess, and is likely not a problem with pip.

Note that there are warnings about the compiler preceding the messages from CUDA. They all seem to indicate that your system is using an unsupported compiler version. You state that your system is derived from Fedora. The Linux Installation Guide for CUDA 11.8 indicates that the following is supported:

Fedora 35: kernel 5.14.10, gcc 11.2.1, glibc 2.34

From the CUDA warning, it appears that your system has gcc 12.x installed. Because of various intricate dependencies, it is generally not possible to mix and match arbitrary software components.

I would suggest first checking what component requirements are imposed by xformers (I have no idea what that is). Once you have worked out which CUDA versions it supports, install a supported CUDA version and the required host software components, as specified in the Linux Installation Guide for that CUDA version.