microsoft / unilm

Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
https://aka.ms/GeneralAI
MIT License
19.08k stars 2.43k forks source link

Kosmos 2.5 for Volta GPU #1587

Open Borobo opened 4 days ago

Borobo commented 4 days ago

I'm on a Volta GPU, I'm aware that flash attention is not compatible with them but I've seen that the kosmos 2.5 requirements.txt contains "xformers", but I haven't seen any implementation in the kosmos 2.5 code. Do you plan to use xformers as a fallback when flash_attn is not installed?