Open thiswillbeyourgithub opened 1 month ago
there are a few complex parts for this.
PS. you might try an older deepspeed version, like 0.13, IIRC this was more compatible at the time.
In linux, you need to add the CUDA development toolkit, or switch to using a CUDA dev image (at least), it probably also needs additional dependencies
Thanks. But does deepspeed only improve the checkpoint loading time or is it faster overall?
It should allow running in lower vram with good performance (but probably not better than fully loaded). At it's core, I think it's essentially efficient layer swap space to ram - I'm not really sure, it may do more than that. deepspeed didn't make any difference at all for me when loading xtts in sufficient vram.
Alright thank you very much for all this clarification. I've decided then not to spend even more time trying. Fish quantization and piper gpu seem a safer bet for lower latency and better speed tradeoff. As far as I'm concerned you can close this. Thanks again!
Hi,
(As per that request) Deepspeed seems to be a library that increases speed for AI related code that support it.
XTTS supports it.
On a non windows computer it seems to be straightforward: just
pip install deepspeed
then use the appropriate XTTS argument. But the issues seem to arise when we're inside a docker container. There's also an issue with deepspeed causing an increase container size, above the threshold allowed by gchr.ioIf you could give pointers to help users try to get deepspeed working on their end it would be awesome! I'm a linux only person. pip install works perfectly outside of docker, but when I tried inside bash of the container I got this error: