Open zhangyuqi-1 opened 2 months ago
https://github.com/InternLM/xtuner/blob/main/requirements/runtime.txt transformers >= 4.43.0 use _flash_attention_forward but not self._flash_attention_forward
Same issue when trying Qwen2.5 sequence parallel and fix it by downgrade transformers to 4.42.4. Will the owner migrate the code to support different versions of transformers?
https://github.com/modelscope/ms-swift/issues/2107#issuecomment-2370042882
xtuner 0.1.23
rank1: File "/data1/zhangyq/miniforge3/envs/py310/lib/python3.10/site-packages/xtuner/model/modules/dispatch/qwen2.py", line 160, in qwen2_attn_forward rank1: attn_output = self._flash_attention_forward( rank1: File "/data1/zhangyq/miniforge3/envs/py310/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1729, in getattr rank1: raise AttributeError(f"'{type(self).name}' object has no attribute '{name}'")