Closed xyfZzz closed 1 month ago
@xyfZzz, you can try searching for libcudart.so.12
using the command find / -name libcudart.so.12
.
This may reveal a path similar to path/to/site-packages/nvidia/cuda_runtime/lib/libcudart.so.12
.
Execute export LD_LIBRARY_PATH=path/to/site-packages/nvidia/cuda_runtime/lib:$LD_LIBRARY_PATH
in your terminal.
Alternatively, you can add this line to your ~/.bashrc
file and then runsource ~/.bashrc
to apply the changes.
Try import vllm_flash_attn
to verify if this work.
libcudart.so.11.0
doesnot work?
but how to use libcudart.so.12
using cuda118
@xyfZzz, you can try searching for
libcudart.so.12
using the commandfind / -name libcudart.so.12
. This may reveal a path similar topath/to/site-packages/nvidia/cuda_runtime/lib/libcudart.so.12
. Executeexport LD_LIBRARY_PATH=path/to/site-packages/nvidia/cuda_runtime/lib:$LD_LIBRARY_PATH
in your terminal. Alternatively, you can add this line to your~/.bashrc
file and then runsource ~/.bashrc
to apply the changes. Tryimport vllm_flash_attn
to verify if this work.
My environment is cuda12
. For cuda118
, maybe you should take a look at https://github.com/vllm-project/vllm/issues/5232 .
It seems that you are not using the cu118 version of vllm-flash-attn. Could you try switching to that version? @xyfZzz
@duanshengliu @ShuaibinLi @fyuan1316 Thank you very much. After I upgraded the cuda version to 12.4, this warning no longer appeared.
Your current environment
🐛 Describe the bug
But I have installed vllm-flash-attn: