huggingface / alignment-handbook

Robust recipes to align language models with human and AI preferences
https://huggingface.co/HuggingFaceH4
Apache License 2.0
4.2k stars 357 forks source link

how to use dpo without flash-attention #91

Open Fu-Dayuan opened 6 months ago

Fu-Dayuan commented 6 months ago

Is there any flash-attention free version?

eryk-mazus commented 5 months ago

As far as I can see, use_flash_attention_2 is set to False by default:

https://github.com/huggingface/alignment-handbook/blob/87cc800498b17432cfb7f5acb5e9a79f15c867fc/src/alignment/configs.py#L140C5-L140C26