-
Is this expected?
I'm on master, commit hash: `59119f4f20b27`
Machine:
Apple M1 Pro, 14-inch, Sonoma 14.3.1
Logs:
```
whisper_init_from_file_with_params_no_state: loading model from 'ggml-…
-
I'm sorry for polluting the GitHub issues with non-bugs, but since that precedent was already set by #1 and there's no Discussions enabled, I thought it may appropriate to share it here.
Laptop CPU…
-
# Prerequisites
Please answer the following questions for yourself before submitting an issue.
- [x] I am running the latest code. Development is very rapid so there are no tagged versions as of…
-
I tried WizardCoder and starcoderplus (ggml q40.bin files) but both are throwing error in linux though when I tried wizardcoder in windows it worked (though it was slow). I am putting the error as bel…
-
How should I convert my model(e.g. .onnx format) to .gguf format and perform inference under the ggml inference framework? How should I implement it step by step?
-
Hi,
Thanks for the simple guide, however, it's not working, anything mismatched?
```sh
./main -m models/ggml-vicuna-13B-1.1-q5_1.bin --repeat_penalty 1.0 --color -i -r "User:" -f prompts/chat…
-
# there is no file named ggml-model-q4_0.bin
llama_model_load: loading model from './models/7B/ggml-model-q4_0.bin' - please wait ...
llama_model_load: failed to open './models/7B/ggml-model-q…
-
One of the major demons I fought while working on https://github.com/saharNooby/rwkv.cpp/pull/74 is ggml's mysterious computation graph work tensor, that is allocated the first time `ggml_graph_comput…
-
I'm still not 100% sure whether to call it llava.cpp or by another name to indicate its future support for other multimodal generation models in the future --maybe multimodal.cpp or lmm.cpp (large mul…
-
Good afternoon,
I was taking a look at https://github.com/ggerganov/ggml/tree/master/examples/gpt-2. There are several ways to obtain ggml compliant GPT2 weights. If I would have a PyTorch-trained …