[x] I reviewed the Discussions, and have a new bug or useful enhancement to share.
Expected Behavior
These two Python scripts should return the same vector.
from langchain.embeddings import LlamaCppEmbeddings
input_size = 5251
llama = LlamaCppEmbeddings(model_path="/home/david_manouchehri/CodeLlama-7b_ggml-model-f16.gguf", n_gpu_layers=1500, n_ctx=input_size, n_batch=input_size)
text = "I am a tomato."
results = llama.embed_documents([text])
print(results)
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
These two Python scripts should return the same vector.
vs.
Current Behavior
Right now, these two scripts return two different vectors. I'm not sure why, as they should be identical.
I'm unsure if this is the same issue as https://github.com/ggerganov/llama.cpp/issues/3287. Still looking into it, just opening this ticket so I and others have something to reference.