Closed ErfolgreichCharismatisch closed 1 month ago
Facing the same loop. Any idea why?
What I found:
I find it odd that everyone pays and uses ChatGPT and basically nobody uses the local models.
Before we continue with this, let's see this video(starts at the right time, just play): https://youtu.be/ywT-5yKDtDg?t=2717
I found a very good alternative as I was also getting the endless loop of doom, not to mention how unbelievably slow it was. Using sentence transformers with HuggingFaceEmbeddings is incredibly fast and works consistently for me. This is, of course, instead of LlamaCppEmbeddings, which seemed to be the culprit.
from langchain.embeddings import HuggingFaceEmbeddings embeddings_model = "sentence-transformers/all-MiniLM-L6-v2" embeddings = HuggingFaceEmbeddings(model_name=embeddings_model)
LlamaCppEmbeddings and HuggingFaceEmbeddings are depreciated now ... any updated example ?
Hello everyone, we have just released version 5, which completely outsources all LLM management to https://github.com/BerriAI/litellm.
If your issue persists, please reopen a new issue using paper-qa>=5
I have the following code for qa with llamacpp and this is what I get, it keeps outputting llama_print_timings, what to make of that?
My code is
This is the output