Open JosefKuchar opened 10 months ago
Thanks for your comments. We recommend you use SDXL model as the diffusion backbone, which would definitely cost no more than 24GB. And kindly note that the cost is positively correlated with the complexity of text prompt.
I had this problem when running the demo (python RPG.py --demo) on 32GB VRAM; running a normal prompt worked (with API key etc.). However, I still don't understand how to use it; I got two people in the output (and that wasn't the prompt). Digging into it :)
感谢您的评论。我们建议您使用 SDXL 模型作为扩散主干,其成本肯定不超过 24GB。请注意,成本与文本提示的复杂性呈正相关。
如何多卡加载
@YangLing0818 Is there a simple way to enable multi-gpu for this?
torch.cuda.OutOfMemoryError: CUDA out of memory. Tried to allocate 7.04 GiB. GPU 0 has a total capacty of 24.00 GiB of which 4.76 GiB is free. Including non-PyTorch memory, this process has 17179869184.00 GiB memory in use. Of the allocated memory 10.53 GiB is allocated by PyTorch, and 7.39 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_CUDA_ALLOC_CONF
I have two 3090s
Ubuntu 22.04.3 LTS (GNU/Linux 5.15.133.1-microsoft-standard-WSL2 x86_64)
Great work!
same here, 24GB GPU (A10) but it fails
nvida-smi says before running demo
there's 0 memory usage.
Ran into oom when running demo, is 24gb not enough?: