huggingface / alignment-handbook

Robust recipes to align language models with human and AI preferences
https://huggingface.co/HuggingFaceH4
Apache License 2.0
4.53k stars 393 forks source link

Downloading latest CUDA version (11.6 or above) for MacOS to use FlashAttention #130

Open shubhamcs162 opened 7 months ago

shubhamcs162 commented 7 months ago

It seems that FlashAttention is only supported on CUDA 11.6 and above. According to https://developer.nvidia.com/cuda-downloads, it seems that the latest version of CUDA (12.3) can't be downloaded for MacOS. I would appreciate some suggestions on getting around this issue. Thank you!