How can I find out why this project didnβt use my Titan RTX (Turing) instead of the smaller RTX 3050 (Ampere 8GB)
My desktop has
- RTX 3050 TI 8GB (Ampere) - id 1
- Titan RTX (Turing) - id 0
The project requires a GPU and appears to have tried to load the model into the 8GB card. How can I tell if it is because it needed a feature not on the Titan?
Loading pipeline components...: 0%| | 0/7 [00:00<?, ?it/s]--- MODELS: Loading Model stabilityai/stable-diffusion-xl-base-1.0 ---
Loading pipeline components...: 29%|βββ | 2/7 [00:00<00:00, 6.67it/s]
Loading pipeline components...: 57%|ββββββ | 4/7 [00:00<00:00, 8.08it/s]
Loading pipeline components...: 71%|ββββββββ | 5/7 [00:00<00:00, 6.95it/s]
Loading pipeline components...: 100%|ββββββββββ| 7/7 [00:00<00:00, 9.52it/s]
Traceback (most recent call last):
File "/usr/lib/python3.10/copy.py", line 231, in _deepcopy_dict
y[deepcopy(key, memo)] = deepcopy(value, memo)
File "/usr/lib/python3.10/copy.py", line 153, in deepcopy
y = copier(memo)
File "/home/workbench/.local/lib/python3.10/site-packages/torch/nn/parameter.py", line 68, in __deepcopy__
self.data.clone(memory_format=torch.preserve_format), self.requires_grad
torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 14.00 MiB. GPU 0 has a total capacity of 8.00 GiB of which 0 bytes is free. Of the allocated memory 7.10 GiB is allocated by PyTorch, and 224.58 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
INFO:httpx:HTTP Request: GET https://checkip.amazonaws.com/ "HTTP/1.1 200 "
INFO:httpx:HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
--- MODELS: Configuring Pipe ---
--- MODELS: Model is ready for inference ---
http://localhost:8000
IMPORTANT: You are using gradio version 4.35.0, however version 4.44.1 is available, please upgrade.
--------
--------