Open josevalim opened 1 year ago
More on attention: https://pytorch.org/blog/flash-decoding/
I'd also suggest FlashAttention-2 and Medusa
Alternative to DPM Solver: https://arxiv.org/abs/2311.05556
I tested SD v1-4 on a GPU using the new lower precision options params_variant: "fp16", type: :bf16
. Here are a couple runs:
Type | Steps | Batch, Images | Time | Memory | Lazy transfers |
---|---|---|---|---|---|
bf16 | 20 | 1, 1 | 0.7s | 4669MiB | No |
bf16 | 20 | 1, 4 | 2.2s | 8769MiB | No |
f32 | 20 | 1, 1 | 1.3s | 8759MiB | No |
f32 | 20 | 1, 4 | 4.3s | 16951MiB | No |
bf16 | 20 | 1, 1 | 3.7s | 6957MiB | Yes |
f32 | 20 | 1, 1 | 8.2s | 13379MiB | Yes |
Note that the reported memory is just the final memory after using preallocate: false
, so it's not ideally reliable. XLA even does memory reservations at compilation time, my guess is that it runs some example operations to pick preferable algorithm or fine tune algorithm parameters. That said, it seems clear that bf16 reduces both memory and time roughly by a factor of 2. Weirdly, lazy transfers seem to bump the memory usage (but it doesn't mean that much memory is required in practice, it's just XLA bumping the reservation, see below).
I experimented with different values of memory_fraction
as an upper limit. For the first entry in the table above:
lazy_transfers: :always
- 3GiB (3.4s)backend_copy
- 4.6GiB (0.7s)preallocate_params: true
it's 6.2GiB (0.7s)So lazy transfers do help a bit, but imply a significant slowdown.
What's interesting though is that preallocate_params
requires more memory than manual backend_copy
. It's even more surprising given that the OOM happens at serving runtime, not during the params preallocation.
preallocate/jit will transfer the data twice, one as arguments, one as return type. So we probably need a new callback/abstraction to make this easier :D
FTR fixed in #317, now preallocate_params: true
effectively does backend_copy
:)
I have added an entry for LCM+Lora, @wtedw may have input here (and we may need to update/release a Axon before). /cc @seanmor5
I think we should update Axon to better support LoRA, I have a draft in place right now but I have to revisit it to make it work as I intend :)
LCM just adapts these nodes in the unet model: https://github.com/wtedw/lorax/blob/main/lib/lorax/lcm.ex#L121-L139 The weights can be found here: https://huggingface.co/latent-consistency/lcm-lora-sdv1-5
For Bumblebee, (if trying to make it compatible w/ most LoRA files in HuggingFace)
If you guys need any PRs, lmk!
Just a heads up that Stability AI just announced Stable Diffusion 3, so that makes us wonder how much effort we should pour into SD vs SDXL vs SD3. It still probably makes sense to support LoRA on Stable Diffusion, because that will require improvements in Axon and elsewhere that we could use for other models, but custom schedulers and token merging is up to debate at the moment.
Checking off attention slicing, it has actually been removed from diffusers docs (https://github.com/huggingface/diffusers/issues/4487) because of flash attention. Either way, the trick is about slicing a dimension and using a while loop, which is similar to flash attention on defn level (as opposed to custom CUDA kernel), and that didn't turn out to be beneficial.
The main part of StableDiffusion is iterative U-Net model pass, which happens for a specified number of timesteps. DeepCache is about reusing some of the intermediate layer outputs across some diffusion iterations, that is outputs expected to change slowly over time.
This technique is not going to reduce memory usage, because we still need to periodically do a uncached model pass. Given that we need to keep the cached intermediate results, it can increase the usage if anything. It can have a significant speedup, assuming we do a fair amount of steps. For SD Turbo or LCM, where we do 1 or at most a few steps, the caching is not applicable.
So this may be something we want to explore in the future, depending on SD3 and other research going forward, but I don't think it's immediately relevant for us now.
A list of ideas to explore: