-
1. when chatting with a model [Hermes-2-Pro-Llama-3-8B-GGUF](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B-GGUF), I get about four questions in, and it becomes extremely slow to generate…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
### What is the issue?
I get this error when I am trying to load this model. Other llama 3.1 models in the Ollama library work great.
(base) PS C:\Users\razva> ollama run CognitiveComputations/dol…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
### Bug Description
In the `run()` method (line 542 in llama_index.core.ingestion.**pipeline.py**) the parameter show_progress is passed to the `run_transformation()` method, but this method doesn'…
-
### Bug Description
I am trying to save my indexVectorStore in Opensearch, but it does not accept the nodes I created with the automerging retriever strategy.
The leaf nodes created with the `Hierar…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
OKUA1 updated
3 weeks ago
-
### Describe the issue as clearly as possible:
When using `models.llamacpp` and creating JSON using a Pydantic model I get an error when generating the first result (see code to reproduce below). I h…
-
### Describe the bug
I have downloaded Hugging Face "meta-llama/Meta-Llama-3.1-8B-Instruct" model to do Q8_0 type quantization using the latest llama.cpp to keep it up-to-date, increase efficiency an…