Open jiapei100 opened 1 year ago
I got the same error too. You should downgrade the commit version of llama.cpp to 41c674161fb2459bdf7806d1eebead15bc5d046e
@c0sogi
Tried.... Not working for me.
➜ llama.cpp git:(master-41c6741)
localGPT still got:
AVX = 1 | AVX2 = 1 | AVX512 = 0 | AVX512_VBMI = 0 | AVX512_VNNI = 0 | FMA = 1 | NEON = 0 | ARM_FMA = 0 | F16C = 1 | FP16_VA = 0 | WASM_SIMD = 0 | BLAS = 1 | SSE3 = 1 | VSX = 0 |
Enter a query: Hi, how are you today?
ggml_new_tensor_impl: not enough space in the context's memory pool (needed 13745376, available 12582912)
[2] 1051032 segmentation fault (core dumped) python run_localGPT.py --device_type cuda
CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 python setup.py bdist_wheel
, I can build out a wheel and have it installed as:CMAKE_ARGS="-DLLAMA_CUBLAS=on" FORCE_CMAKE=1 pip install -e .
, I'll build out the following, without llama_cpp.Unbelievable, it resorts to llama.cpp under the folder vendor. from where it's been built.
LLAMA_ASSERT: ....../llama-cpp-python/vendor/llama.cpp/llama.cpp:1800: !!kv_self.ctx
.Okay... Can anybody please tell me how to build llama-cpp-python from source and have it successfully installed in Release mode?