Open li126com opened 10 months ago
Implement deterministic backward (thanks to Meituan) This commit is not implemented in the mha.py. Parameter calling should be added to both FlashSelfAttention and FlashCrossAttention.
Thanks, do you want to add a PR?
sure, https://github.com/Dao-AILab/flash-attention/pull/748
Implement deterministic backward (thanks to Meituan) This commit is not implemented in the mha.py. Parameter calling should be added to both FlashSelfAttention and FlashCrossAttention.