Open Crazy-LittleBoy opened 9 months ago
Yup, it's mentioned in the README
FlashAttention-2 currently supports:
Ampere, Ada, or Hopper GPUs (e.g., A100, RTX 3090, RTX 4090, H100). Support for Turing GPUs (T4, RTX 2080) is coming soon, please use FlashAttention 1.x for Turing GPUs for now.
Support for Turing GPUs (T4, RTX 2080) is coming soon. Looking forward to it. @tridao
Unfortunately I've had no bandwidth to work on this. We welcome contributions.
flash attention 1 support turing, but flash attention 2 not ?