Closed batrlatom closed 9 months ago
I quickly tested this, it seems to be non-functional on older versions of llamacpp, but works on latest. I suppose you could try pull the latest llamacpp files to fix this. Otherwise, @jhen0409 will have to bump the llamacpp version.
I will try... thanks
@Vali-98 Tried to update submodule to latest version, but still the same problem ... you can take a look at: https://github.com/batrlatom/llama.rn . for reference, I installed it via git using
npm install git+https://github.com/batrlatom/llama.rn
Supposedly its was added on this commit: https://github.com/ggerganov/llama.cpp/commit/d6bd4d46ddb6926087c11e0f6633ab1c81da58c3
Checking the npm install that you suggested, it seems to pull the incorrect version of llamacpp, these lines are missing:
// optional bias tensors, present in Stable LM 2 1.6B
layer.bq = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_Q, "bias", i), {n_embd}, false);
layer.bk = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_K, "bias", i), {n_embd_gqa}, false);
layer.bv = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_V, "bias", i), {n_embd_gqa}, false);
Something seems to have gone wrong with your compilation. It seems to be in your repo, but it doesn't exist when using npm install.
ok, it's important that it works for you so I know that I need to find a problem on my side. Thanks!
solved, needed to run bash scripts/bootstrap.sh
to update needed files
Hello,
I've been working on getting the stablelm-2-zephyr-1_6b-Q8_0.gguf operational (link: https://huggingface.co/spaces/stabilityai/stablelm-2-1_6b-zephyr), especially since the 3B version seems to function quite well. However, I'm encountering an issue with the 1.6B version where it fails to initialize the context. Currently, I'm using the latest version of your master branch to compile the library. Is there a straightforward modification I can make on my end to resolve this?
from logs:
01-29 22:50:05.365 3017 20732 E RNLLAMA_LOG_ANDROID: llama_model_load: error loading model: done_getting_tensors: wrong number of tensors; expected 340, got 268
Thank you.