Open Tortoise17 opened 2 months ago
Flash attention 1.x supports Turing, Flash attention 2.x doesn't support Turing as of right now.
@Carnyzzle Thank you. I downgraded to flash_attn 1.xx and still same error. If you can specifically mention which version can get it resolved would be a great help
I am facing this error
RuntimeError: FlashAttention only supports Ampere GPUs or newer.
while architecture is Turing. Is there any tip to resolve it? GPU is NVIDIA T1000.
kindly help.