I ran python scripts/txt2img.py --prompt "a photograph of an astronaut riding a horse" --plms --n_samples=1 --precision=full
Error message:
RuntimeError: CUDA out of memory. Tried to allocate 1024.00 MiB (GPU 0; 6.00 GiB total capacity; 5.06 GiB already allocated; 0 bytes free; 5.17 GiB reserved in total by PyTorch) If reserved memory is
>> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
Is there a way to have PyTorch not allocate almost all of my VRAM? I'm guessing that's the problem here.
I ran
python scripts/txt2img.py --prompt "a photograph of an astronaut riding a horse" --plms --n_samples=1 --precision=full
Error message:
Is there a way to have PyTorch not allocate almost all of my VRAM? I'm guessing that's the problem here.