-
**Describe the bug**
Running the ilab generate command results in an empty dataset being generated for the knowledge I'm trying to add:
```
ilab generate --model-family merlinite --sdg-scale-fa…
-
I have been staging some updates testing the tgi-gaudi software with llama 405B fp8, i am waiting for habana optimum to approve the PR, and then I will submit a pr for huggingface/tgi_gaudi and will s…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
Version:
llama_cpp_dart 0.0.6
llama.cpp tag: b2277
logcat:
```bash
02-28 00:21:29.079 5839 8926 E flutter : [ERROR:flutter/runtime/dart_isolate.cc(1107)] Unhandled exception:
02-28 00:21:2…
-
**Describe the bug**
The model correctly asks to use a tool (the message #2 below), but when the response is sent back to the inference server with the result of the tool, the response message …
feloy updated
4 weeks ago
-
I noticed that chatml is the only format supported right now in llama file, while llama.cpp already supports multiple formats.
Will llama file sync up with llama.cpp and support those other formats …
-
**LocalAI version:**
Latest
**Environment, CPU architecture, OS, and Version:**
EC-2
**Describe the bug**
Getting the grpc connection error when running using cuda12 image. But when runni…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [ Yes] I am running the latest code. Development is very rapid so there are no tagged versions as…
-
I am trying to execute the following script:
1. from llama_cpp import Llama
2. llm = Llama(model_path="~/llama-2-7b.ggmlv3.q8_0.bin", n_gqa=8)
3. output = llm("Q: Name the planets in the solar sy…
-
### What happened?
Going through the manual - SYCL I mean. Everything compiles okay. Running it always thows an error. Can't make it work. OS used: Linux Gentoo. P.S. docker doesn't work either. P.P.…