-
Need to add functional testing for:
(pending #247)
- [x] summarizer
- [x] codegen
- [x] rag
- [x] whisper model-server
- [ ] llamacpp-python-cuda
- [ ] llamacpp-python-vulkan
- [ ] audio-to-…
-
I use this instruction
https://continue.dev/docs/reference/Model%20Providers/llamacpp
But when I try to connect, VSCode (Continue) writes an error:
**Continue error: HTTP 404 Not Found from http://…
-
I noticed this topic: Pipeline parallelism improves batch processing performance when using multiple GPUs, [https://github.com/ggerganov/llama.cpp/pull/6017](url)
I'm not well versed in the topic, …
-
when changing from nitro.exe rest service to cortex-cpp.exe some breaking changes are introduced
1. load and unload models with
before: /inferences/llamacpp/[un]loadmodel has changed to
after: /…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
r3v1 updated
6 months ago
-
hellaswag error - llamacpp build main: build = 2673 (04fbc5f2)
````
build\bin\perplexity.exe --model models/new3/WizardLM-2-7B.Q8_0.gguf --threads 30 --ctx-size 0 -ngl 99 --hellaswag --hellaswag-…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [X ] I am running the latest code. Development is very rapid so there are no tagged versions as o…
-
thanks for Tencent's amazing HunyuanDiT(it's very cool in "Hunyuan Assistant(aka 混元助手)" as a Wechat Mini App(aka Wechat Xiaochengxu or 微信小程序)).
1. could you provide tutorial to illustrate how to ru…
-
https://outlines-dev.github.io/outlines/reference/models/llamacpp/
```
from llama_cpp import Llama
from outlines import models
llm = Llama.from_pretrained(
repo_id="Qwen/Qwen1.5-0.5B-Chat…
-
### Have you searched for similar requests?
Yes
### Is your feature request related to a problem? If so, please describe.
Currently it is not possible to use the API key with the llamacpp text back…