-
### What is the issue?
My cards is w7900, and rocm driver is 6.3 , I found the llama-cpp server started by Ollama always without -fa flag.
I check the code , found :
…
-
Hi Team,
I am already using LMStudio and OLLAMA for model deplyments. Given this model is LMCPP compatible and uses that. How can this model be deplyment, hosted and used with LMStudio or OLLAMA. It …
-
# Expected Behavior
I tried to install llama via poetry and it didnt work
# Current Behavior
it just prompted some information that i dont understand, tried checking, asked for help and it …
-
## Overview
- Intel's Lunar Lake is releasing soon, which has CPU, NPU and iGPU in a single chip
## Tasklist
- [x] https://github.com/janhq/cortex.cpp/issues/677
- [x] https://github.com/janhq/cort…
-
Hi, I wanted to try the model code-7b, but I got this error :
```
llama-gpt-llama-gpt-ui-1 | [INFO wait] Host [llama-gpt-api:8000] not yet available...
llama-gpt-llama-gpt-api-1 | /usr/local/…
-
Hi, do you think I can just drop newer versions of llama.cpp, ggml.h ggml.c etc into src to keep this up to date with llama.cpp? Or is there more too it?
-
This reports mistral.rs as being faster than llama.cpp: https://github.com/EricLBuehler/mistral.rs/discussions/612
But I'm seeing much slower speeds for the same prompt/settings.
Mistral.rs
``…
-
**Is your feature request related to a problem? Please describe.**
When editing the beginning of a long file, prompt evaluation takes a lot of time.
Reason for that - in `Additional context`
Curr…
-
I'm attempting to install llama-cpp-python under the tensorflow-gpu docker image (nightly build) . When I attempt to do so, I get the following error messages.
````
root@a1f1e127514b:/tf# CMAKE_A…
-
Hi thanks for the lib! I want to use some embedding models (arch is bert) from hf hub. I have tried gguf, but the converter says bert arch cannot be converted to that. I have also tried directly have …