-
### Describe the bug
Attempting to load a model after running the update-wizard-macos today (the version from a day or two ago worked fine) fails with the stack trace log included below.
Notabl…
-
### What happened?
When I tried to quantize using the following command, I got the following error. Do you know the cause?
`py convert-hf-to-gguf.py --outtype f16 F:/models/Llama-3-Lumimaid-70B-…
-
When excution example workflow
![image](https://github.com/gokayfem/ComfyUI_VLM_nodes/assets/142652112/e9df99a2-3efe-4137-91c9-93c5cf44fb62)
> !!! Exception during processing!!! could not broadcas…
-
I'm trying to quantize llama2 70B using [AutoFP8 with calibration samples from ultrachat](https://github.com/neuralmagic/AutoFP8/blob/147fa4d9e1a90ef8a93f96fc7d9c33056ddc017a/example_dataset.py) in a …
-
Hi,
On ubuntu 22.04, It just stucks at generating output for hours:
# Import the Llama class of llama-cpp-python and the LlamaCppPythonProvider of llama-cpp-agent
from llama_cpp import Llama
fr…
-
**Describe the bug**
Using v0.2.146, installation works fine, but when i finished to create the integration, i got a "failed to configure" message.
**Expected behavior**
Integration shall s…
pbn42 updated
5 hours ago
-
When I try to use your llama-cpp-server-cuda:main I get this error
ghcr.io/allenporter/llama-cpp-server-cuda:main
docker: Error response from daemon: failed to create task for container: failed to…
-
### Describe the problem
I would like if the collections class had a built-in function for using a local model and llamacpp to embed the documents
### Describe the proposed solution
Since I have no…
-
### What happened?
I've compiled the latest build of `llama-server` (version b3205) using the method recommended in the docs.
The recent renaming of `server` to `llama-server` has tempted me to us…
-
### What is the issue?
### Issue with Getting Model over Ollama AP api/models after Creating Model in Ollama
#### Steps to Reproduce
1. Execute the following command to merge GGUF files:
e…