Closed jamesbondzhou closed 1 month ago
推理使用的模型是'OpenGVLab/Mini-InternVL-Chat-4B-V1-5'
Hello, the installation of Flash Attention might not have been successful. You can try uninstalling Flash Attention first and then run it again, or follow the official instructions for Flash Attention to reinstall it: https://github.com/Dao-AILab/flash-attention.
Hello, the installation of Flash Attention might not have been successful. You can try uninstalling Flash Attention first and then run it again, or follow the official instructions for Flash Attention to reinstall it: https://github.com/Dao-AILab/flash-attention.
YES! Uninstalling Flash Attention ,it works! 3KS Bro
Checklist
Describe the bug
按照提示装好环境后,执行 model=AutoModel.from_pretrained(path,torch_dtype=torch.float16,low_cpu_mem_usage=True,trust_remote_code=True,device_map='auto').eval().cuda()报错,如下:
File /data/anaconda3/envs/videochat_new/lib/python3.9/site-packages/transformers/models/llama/modeling_llama.py:55 54 if is_flash_attn_2_available(): ---> 55 from flash_attn import flash_attn_func, flash_attn_varlen_func 56 from flash_attn.bert_padding import index_first_axis, pad_input, unpad_input # noqa
File /data/anaconda3/envs/videochat_new/lib/python3.9/site-packages/flash_attn/init.py:3 1 version = "2.3.6" ----> 3 from flash_attn.flash_attn_interface import ( 4 flash_attn_func, 5 flash_attn_kvpacked_func, 6 flash_attn_qkvpacked_func, 7 flash_attn_varlen_func, 8 flash_attn_varlen_kvpacked_func, 9 flash_attn_varlen_qkvpacked_func, 10 flash_attn_with_kvcache, 11 )
File /data/anaconda3/envs/videochat_new/lib/python3.9/site-packages/flash_attn/flash_attn_interface.py:10 8 # isort: off 9 # We need to import the CUDA kernels after importing torch ---> 10 import flash_attn_2_cuda as flash_attn_cuda 12 # isort: on
ImportError: libtorch_cuda_cpp.so: cannot open shared object file: No such file or directory
Reproduction
model=AutoModel.from_pretrained(path,torch_dtype=torch.float16,low_cpu_mem_usage=True,trust_remote_code=True,device_map='auto').eval().cuda()
Environment
Error traceback
No response