-
### What happened?
I tried to run the tinyllama-1.1b model on a [OnePlus CPH2573](https://vulkan.gpuinfo.org/listreports.php?devicename=OnePlus+CPH2573&platform=android) (with Adreno™ 750). It works …
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.md)…
-
## Description
This issue tracks the process of facilitating this integration and ensures our repository is ready for incorporation into LlamaIndex.
## Objectives
- [ ] Evaluate the compatibi…
-
Running the `inflation.py` example from the rep. I am expecting it calls the custom tool for `get_ticker_data` function, which is defined at the folder `custom_tools` by `ticker_data.py`. However, ba…
-
When I run `poetry run generate` I get
```FileNotFoundError: [Errno 2] No such file or directory: 'config/loaders.yaml'```
Where can I find information on this file?
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ ] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
### Description
i am trying the below code but getting multiple errors
```
using LLama.Common;
using LLamaSharp.KernelMemory;
using Microsoft.KernelMemory;
using Microsoft.KernelMemory.C…
-
Installed from conda environment with pip
**Version: '0.2.69'**
The code is as follow:
```
llm = Llama(
model_path="/data/codelama-2024-02/CodeLlama-7b-Python/ggml-model-f16.gguf",
…
-
### Bug Description
The retriever on graph index makes an assertion statement false. The assertion is the the following:
assert isinstance(node, BaseNode) or node is None
That is present inside t…
-
### What happened?
I am trying to run inference on RPC example. When running the llama-cli with rpc feature over a single rpc-server on localhost, the inference throughput is only 1.9 tok/sec for lla…