Open couragelfyang opened 10 months ago
Try with cuda >= 12
Similar issues with 11.xx https://github.com/huggingface/candle/issues/353
I think it's a bug with last minor update or so it seems, because i tried compiling with llama-cpp-python==0.2.27
and it worked normally.
Full command for building with CUDA support: FORCE_CMAKE=1 CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python==0.2.27
@MrJefter latest minor update only updated llama.cpp so unless something broke there nothing here should've changed.
@abetlen anyway, downgrade to 0.2.27 fixes this issue
We should invent/use a way to only download stable versions when doing pip install X
rather than installing bleeding edge. Minor versions should be only for the repo's devs really as they tend to break.
The error message clearly points to a llama.cpp
issue...
vendor/llama.cpp/ggml-cuda.cu(626): error: identifier "__hmax2" is undefined
I'm trying to install llama-cpp-python through
CMAKE_ARGS="-DLLAMA_CUBLAS=on" pip install llama-cpp-python
but met this error. Any suggestion?