-
Hi, awesome project!
I'm on the doorstep of my first query, but I'm stuck.
This is the Ollama server API endpoint:
```bash
curl http://10.4.0.100:33821/api/version
{"version":"0.4.2"}
```
T…
-
Hello,Sorry to bother you.
T tested the PPLs of llama.cpp and T-MAC is abnormal, which values are 110682 and 53515, so big. But we know that the normal value should be very small. So then I try to t…
-
**LocalAI version:**
localai/localai:latest-gpu-nvidia-cuda-12 : SHA ff0b3e63d517
(Also occurs on v2.22.1 container image)
**Environment, CPU architecture, OS, and Version:**
Linux server 6.8.0-…
-
- [LlamaCpp](https://python.langchain.com/docs/integrations/llms/llamacpp)
- https://github.com/netdur/llama_cpp_dart
- Tool calling:
* https://github.com/ggerganov/llama.cpp/issues/5588
* ht…
-
Tracker for: https://github.com/ggerganov/llama.cpp/discussions/5138 and also ROCm
- [x] Vulkan: https://github.com/mudler/LocalAI/pull/2648 (upstream https://github.com/ggerganov/llama.cpp/pull/20…
-
Hi,
Ternary quantization has become popular and has demonstrated computational speedups and power reductions, as demonstrated in works like llama.cpp and [bitnet.cpp](https://github.com/microsoft/B…
-
After running `.\tabby.exe serve --model StarCoder-1B --chat-model Qwen2-1.5B-Instruct` after downloading the models, the llama-server.exe crashes.
Tabby
tabby v0.18.0 and tabby v0.19.0-rc.1
tabb…
-
### Motivation
Many llama.cpp users are requesting this so far. Ollama is one of the interfaces of llama.cpp and it is quite popular. Implementing it will significantly accelerate InterVL adoption an…
-
### What is the issue?
I followed the instructions for building on mac [here](https://github.com/ollama/ollama/blob/main/docs/development.md#macos), but this failed at the `go generate` step.
Runn…
-