Closed zarif98sjs closed 2 weeks ago
flash_attention_2 is not easy to use, maybe you can try not set attn_implementation parameter, i.e. in finetune.py
model = Qwen2VLForConditionalGeneration.from_pretrained(
"Qwen/Qwen2-VL-2B-Instruct", torch_dtype=torch.bfloat16, attn_implementation="flash_attention_2", device_map="auto"
)
--->
model = Qwen2VLForConditionalGeneration.from_pretrained(
"Qwen/Qwen2-VL-2B-Instruct", torch_dtype=torch.bfloat16, device_map="auto"
)
even without that it needs flash_attention_2
when importing Qwen2VLForConditionalGeneration
.
Anyway I was able to solve the issue by reinstalling flash attention.
even without that it needs
flash_attention_2
when importingQwen2VLForConditionalGeneration
.Anyway I was able to solve the issue by reinstalling flash attention.
cool
I installed everything from
requirements.txt
, but still get this error when I runfinetune.py
: