-
### System Info / 系統信息
cuda 11.8
llama-cpp-python 0.2.55
python 3.10
windows 10
### Running Xinference with Docker? / 是否使用 Docker 运行 Xinfernece?
- [ ] docker / docker
- [X] pip install …
-
### What is the issue?
It's again the https://github.com/ollama/ollama/issues/6011 issue.
**The issue is with embedding call with the model converted using convert_hf_to_gguf.py.**
litellm.ll…
-
### What happened?
Hi there.
I am trying to use the `np` parameter to serve multiple requests in parallel. However, the generated tokens are garbled when I set the `np` parameter to a relatively lar…
-
### How are you running AnythingLLM?
Docker (local)
### What happened?
Docker sees my models. I start chatting in my workspace, and then I get an error "Failed to load model"
```
anythingllm |…
-
| --- | --- |
| Bugzilla Link | [546504](https://bugs.eclipse.org/bugs/show_bug.cgi?id=546504) |
| Status | NEW |
| Importance | P3 normal |
| Reported | Apr 17, 2019 05:26 EDT |
| Modified | Feb…
-
Trying to quantise some flux models to lower the vram needs and I get that error.
```
(venv) C:\AI\llama.cpp\build>bin\Debug\llama-quantize.exe "C:\AI\ComfyUI_windows_portable\ComfyUI\models\chec…
-
### What happened?
Hi there.
My llama-server can work well with the following command:
```bash
/llama.cpp-b3985/build_gpu/bin/llama-server -m ../artifact/models/Mistral-7B-Instruct-v0.3.Q4_1.g…
-
Model placed in brushnet folder, loader cannot read
-
`Actions/SynchronizeAction.php` uses the `Spatie\TranslationLoader\LanguageLine` instead of the model defined in `config/translation-loader.php`
-
### HA core version
2024.11.3
### Current integration version
0.6.5
### Last known working integration version
None
### Device type and model
Front Loader Washer Dryer Smart Wi-Fi…