LLaMA-Omni is a low-latency and high-quality end-to-end speech interaction model built upon Llama-3.1-8B-Instruct, aiming to achieve speech capabilities at the GPT-4o level.
2024-09-27 14:47:10 | ERROR | stderr | File "anaconda3/envs/llama-omni/lib/python3.10/site-packages/transformers/modeling_flash_attention_utils.py", line 180, in _flash_attention_forward
2024-09-27 14:47:10 | ERROR | stderr | _flash_supports_window_size and sliding_window is not None and key_states.shape[1] > sliding_window
2024-09-27 14:47:10 | ERROR | stderr | NameError: name '_flash_supports_window_size' is not defined
2024-09-27 14:47:10 | ERROR | stderr | File "anaconda3/envs/llama-omni/lib/python3.10/site-packages/transformers/modeling_flash_attention_utils.py", line 180, in _flash_attention_forward 2024-09-27 14:47:10 | ERROR | stderr | _flash_supports_window_size and sliding_window is not None and key_states.shape[1] > sliding_window 2024-09-27 14:47:10 | ERROR | stderr | NameError: name '_flash_supports_window_size' is not defined
transformers 4.43.4