-
Can you please add local LLM support, please?
Ollama support will be nice too.
Thank you.
-
Hi there.
Very nice project!
Would it be possible to use an OpenAi compatible APi endpoint with a local LLM through [LM Studio](https://lmstudio.ai/) or [text-generation-webui](https://github.co…
-
I have this GPU: AMD RADEON RX 7900 XTX
ramalama pull "quay.io/ramalama/rocm:latest"
When I try a model, there is always a crash.
e.g.
$ ramalama run llama3.2
"llama-cli killed by SIGSEGV"
cmdline:
…
-
### What is the issue?
`(.venv) [root@bastion ollama]# python llm/llama.cpp/convert-hf-to-gguf.py ./model --outtype f16 --outfile converted.bin
INFO:hf-to-gguf:Loading model: model
INFO:gguf.gguf_…
-
### 起始日期 | Start Date
_No response_
### 实现PR | Implementation PR
_No response_
### 相关Issues | Reference Issues
_No response_
### 摘要 | Summary
need llama.cpp support video by server mode
### 基本…
-
### System Info
```
(zt) root@autodl-container-7071118252-7032359d:~/test/PiPPy/examples/llama# transformers-cli env
Copy-and-paste the text below in your GitHub issue and FILL OUT the two last p…
-
Fixit@DAD MINGW64 ~/LlamaGPTJ-chat
$ mkdir build
(myenv)
Fixit@DAD MINGW64 ~/LlamaGPTJ-chat
$ cd build
(myenv)
Fixit@DAD MINGW64 ~/LlamaGPTJ-chat/build
$ mkdir models
(myenv)
Fixit@DAD MINGW6…
-
**Is your feature request related to a problem? Please describe.**
When running multiple distributed workers, if I have to change or restart the service on a worker, I have to bring the entire …
-
# Expected Behavior
I have a machine with and AMD GPU (Radeon RX 7900 XT). I tried to install this library as written in the README by running
`CMAKE_ARGS="-DLLAMA_HIPBLAS=on" FORCE_CMAKE=1 pip …
-
```
error occurred: Command ZERO_AR_DATE="1" "ar" "cq" "/home/tc-wolf/rust-llama.cpp/target/release/build/llama_cpp_rs-75252caa56296e09/out/libbinding.a" "/home/tc-wolf/rust-llama.cpp/target/releas…