Dao-AILab / flash-attention

Fast and memory-efficient exact attention
BSD 3-Clause "New" or "Revised" License
13.1k stars 1.18k forks source link

Unable to build wheel of flash_attn #1007

Open Zer0TheObserver opened 2 months ago

Zer0TheObserver commented 2 months ago

I'm using CUDA 12.3.r12.3 OS: Windows11 21H2

The process stock in: (from sympy->torch->flash_attn) (1.3.0) Using cached einops-0.8.0-py3-none-any.whl (43 kB) Building wheels for collected packages: flash_attn Building wheel for flash_attn (setup.py) I wait for few hours.

Sayli2000 commented 1 month ago

I'm using CUDA 12.3.r12.3 OS: Windows11 21H2

The process stock in: (from sympy->torch->flash_attn) (1.3.0) Using cached einops-0.8.0-py3-none-any.whl (43 kB) Building wheels for collected packages: flash_attn Building wheel for flash_attn (setup.py) I wait for few hours.

same problem did you manage to solve it?