openai / consistency_models

Official repo for consistency models.
MIT License
6.08k stars 411 forks source link

implement a non-Cuda flash attention module #33

Open by321 opened 1 year ago

by321 commented 1 year ago

The current flash attention module by Hazy Research is Cuda-only, so this limits this repo to Cuda-only too. I suggest writing a separate flash attention module for computers without Nvidia video card. This current module can still be used if Nvidia card is present.

A couple of people have raised this issue with Hazy Research, but they said they're focused on Cuda only, and are not interested in writing a non-Cuda version.

ayushtues commented 1 year ago

Added a pr #37 to avoid using flash Attention