-
In the "The Llama 3 Herd of Models" paper, FFN dimension for the 8B, 70B and 405B models are stated as 6,144, 12,288 and 20,480. I would have expected the parameter count to stay the same as llama 3 w…
-
The official llama 3 70b instruct repo has updated the eos token
"eos_token": "",
Yet when using this library and using that eos token, no output is outputted because it used the old eos t…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
Latest LLaMA-Factory repo 12Septr2024 forces to use Torch 2.4 hence is clashing with Unsloth/XFormers
##…
-
llama.cpp running in server mode, how to use this? any documentation on usage?
-
### Discussed in https://github.com/ggerganov/llama.cpp/discussions/9960
Originally posted by **SteelPh0enix** October 20, 2024
I've been using llama.cpp w/ ROCm 6.1.2 on latest Windows 11 for…
-
Hi, Thanks for your wonderful work.
I am struggling using my lora tuned model.
I conducted following steps
1. finetuning with lora
- Undi95/Meta-Llama-3-8B-Instruct-hf model base
- llama3 …
-
-
Hi,
Thanks for your great work.
I try to reproduce the results of offline dpo and offline simpo and I found the reproduced resltus are better the results in the paper. For example, for the resul…
-
from llama_parse import LlamaParse
--------------------------------------------------------------------------
ImportError Traceback (most recent call last)
Cell In…
-
### What happened?
Hi! Thanks for your efforts for contributing such a great framework!
I am working on deploying a custom service on my PC and learning to make llama.cpp produce structured output v…