It seems that FlashAttention is only supported on CUDA 11.6 and above. According to https://developer.nvidia.com/cuda-downloads, it seems that the latest version of CUDA (12.3) can't be downloaded for MacOS. I would appreciate some suggestions on getting around this issue. Thank you!
It seems that FlashAttention is only supported on CUDA 11.6 and above. According to https://developer.nvidia.com/cuda-downloads, it seems that the latest version of CUDA (12.3) can't be downloaded for MacOS. I would appreciate some suggestions on getting around this issue. Thank you!