Closed PauloIVM closed 5 months ago
Hi @PauloIVM I'll release version 3.0 of this binding very soon, which will probably fix this problem. If you want you can have an early look at the v3.0 branch. It's working on my Linux machine, I'm just struggling to get a CI GitHub workflow running, which I want to finish before merging. The API around ModelParameters
and InferenceParameters
slightly changed, have a look at src/test/java/examples
in the branch.
Thanks so much! I'll take a look at this branch :smile:
I just released version 3.0 which should hopefully solve your problems. Feel free to re-open if you still have issues.
Hello. I'm trying to run this lib on an Ubuntu OS, using GPU acceleration with an Nvidia 1660 Super.
I was able to run the java-llama.cpp lib with a custom llama.cpp, when I built the llama.cpp like this:
Naturally, as I didn't pass the argument "-DLLAMA_CUDA=ON" (or "-DLLAMA_CUBLAS=ON", I tried too with b1645 release), GPU acceleration was not used. But it worked correctly pointing to llama.cpp custom.
If I just rebuild llama.cpp according to the script below, then the code with the java-llama lib starts to break:
I even tried to compile llama.cpp also with this release, which is the one used here by java-llama.cpp, and use
-DLLAMA_CUBLAS=ON
instead ofDLLAMA_CUBLAS=ON
. But the same error remains.So this is my code. It's just a test code, very similar to the example in the jav-llama readme:
And, when using a llama.cpp buited with
-DLLAMA_CUBLAS=ON
, I got this error below (and the code runs without my custom llama, using the default llama.cpp of this java-llama project, without GPU acceleration):If I simply rebuild llama.cpp without the
-DLLAMA_CUBLAS=ON
argument, this error disappears and the code starts running using my custom lib.Any idea why this error occurs? And how to solve it?