-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
pily1 updated
2 months ago
-
When I create `conda` environment using these steps:
```
conda create --name gguf-to-torch python=3.12 -y
conda activate gguf-to-torch
conda install pytorch torchvision torchaudio pytorch-cuda…
-
### 起始日期 | Start Date
_No response_
### 实现PR | Implementation PR
_No response_
### 相关Issues | Reference Issues
_No response_
### 摘要 | Summary
need llama.cpp support video by server mode
### 基本…
-
**The bug**
When using roles such as `with assistant():` combined with `stream()`, roles are not obeyed and all output is being squished into the last role...
**To Reproduce**
Give a full worki…
-
### Cortex version
v172
### Describe the Bug
App shell conflicts between `cortex-nightly` and `cortex-stable`:
when I load a model on one version, the model is loaded on both `cortex` and `corte…
-
Could you provide a script to convert a model from the Lit-LLaMA format to the original format, so that it can be used in llamacpp? The Lit-LLaMA format is not supported by llamacpp.
The `/scripts…
-
The following code appears to load the llamacpp model properly, but it just ramps up the CPU load and hangs for hours if allowed.
If service_context=service_context is removed from GPTSimpleVectorInd…
-
Gemma models that have been quantized using Llamacpp are not working. Please look into the issue
error
"llama.cpp error: 'create_tensor: tensor 'output.weight' not found'"
I will open a issue…
-
-edit- found out more about the issue and added another comment as it affects transformers as well.
**Describe the bug**
This issue seems to be related to using HF model loaders. I don't know wha…
-
Let's start looking at what's happening over here https://github.com/ggerganov/ggml/tree/master/examples/mnist down to the flow of PRs and the kinds of discussions folks are having there (e.g. https:/…