Closed dachenlian closed 1 week ago
I copy pasted gemma2 forward code and monkey patched with our FLCE layer. Seems that it was not removed yet. I will update after they make the change. (maybe they should but forget to do it)
By the way, I have seen you around when i was in NTU's gym. You lifted massive weight XD
Small world 😘 It's unfortunate we didn't hang out 🥺 Oh, and thank you for this amazing work!
https://github.com/linkedin/Liger-Kernel/blob/81d98ea895255a44a0c787c7afa0ab7c34e32884/src/liger_kernel/transformers/model/gemma2.py#L61
Is this warning related to FlashAttention being previously unable to support Gemma2's softcapping? If so, it was fixed in v2.6.