-
### What is the issue?
I successfully converted jina-embeddings v2 base zh to gguf through llama. cpp and imported it into llama。
Here is my Modelfile
> root@buaa-KVM:~/1T/ollama/Jina-AI-embeddin…
-
How to disable this automatic behavior? And if it's not possible yet, can we get a --flag for it?
`
llama_tokenize_internal: Added a BOS token to the prompt as specified by the model but the promp…
-
Mistral AI just dropped Pixtral, their 12b model with vision support.
- https://github.com/mistralai/mistral-common/releases/tag/v1.4.0
- https://www.reddit.com/r/LocalLLaMA/comments/1fe3x1z/mistr…
-
Im toying around with the llama-cpp-python[server] and am running into an odd issue regarding the default port and the port argument when running it from the terminal through SSH into a Ubuntu Linux s…
-
Hi, its a wonderful repository, I have a doubt. I'm new to this.. how did you pretrain the llama2 base model. because malayalam is not trained in the base model right ?, its only trained on english to…
-
Hi, I wanted to try the model code-7b, but I got this error :
```
llama-gpt-llama-gpt-ui-1 | [INFO wait] Host [llama-gpt-api:8000] not yet available...
llama-gpt-llama-gpt-api-1 | /usr/local/…
-
在convert_hf_to_gguf.py文件中,转换MiniCPM模型的时候,如下类override了modify_tensors,并且只转换了q_proj.weight和k_proj.weight,请问为什么需要转换呢?或者如注释所说“HF models permute some of the tensors, so we need to undo that”,HF model是在那里做了这…
-
When I try to use your llama-cpp-server-cuda:main I get this error
ghcr.io/allenporter/llama-cpp-server-cuda:main
docker: Error response from daemon: failed to create task for container: failed to…
-
### Describe the bug
Model URL:
https://huggingface.co/bartowski/Hubble-4B-v1-GGUF/discussions/1
llama_model_loader: - kv 26: tokenizer.ggml.merges arr[str,280147] = ["Ġ Ġ"…
-
Hi, do you think I can just drop newer versions of llama.cpp, ggml.h ggml.c etc into src to keep this up to date with llama.cpp? Or is there more too it?