Closed Maxon081102 closed 2 weeks ago
What does quality dropping mean?
If I understand right, you're comparing a custom embedding model to Jini, and the custom embedding model is worse than Jini. If that is true, it sounds like a model issue, not llama.cpp.
What does quality dropping mean?
If I understand right, you're comparing a custom embedding model to Jini, and the custom embedding model is worse than Jini. If that is true, it sounds like a model issue, not llama.cpp.
This means that when I convert a model to gguf the quality drops on metrics such as MRR and others
No, I use the same model via transformers and llama.cpp, and for some reason the embeddings of llama.cpp are very different, although I converted everything according to the tutorial
This issue was closed because it has been inactive for 14 days since being marked as stale.
What happened?
I tried to use the jini model with mean pooling, sbert like in tutorial and my custom bert with gpt2 tokenizer and mean pooling, and each time the quality dropped,and the average cosine distance of the embeddings was from 0.7 to 0.9, but when i do everything according to the tutorial (I used arctic model for getting results) cosine distance is 0.999
It seems that the quality should not fall. Tell me please, what is wrong? Maybe there are problems with using an unusual bert models?
I converted the models using linux, running it on windows with ollama
A Simple file.py for getting results
Name and Version
llama.cpp: build: 3771 (acb2c32c) with cc (Ubuntu 11.4.0-1ubuntu1~22.04) 11.4.0 for x86_64-linux-gnu
ollama version is 0.1.41
What operating system are you seeing the problem on?
No response
Relevant log output