-
### Model description
Here is the model description
> gte-Qwen1.5-7B-instruct is the latest addition to the gte embedding family. This model has been engineered starting from the [Qwen1.5-7B](https:…
-
Specs: rtx 3060ti w/ 8gb vram, r7 5700x, 32gb ram
main says
`main: build = 2769 (8843a98c)
main: built with cc (Debian 12.2.0-14) 12.2.0 for x86_64-linux-gnu`
make says
`GNU Make 4.3
Built…
-
Hi Team,
It is amazing handbook. In the continued pre-training script (`run_cpt.py`), I saw that it is not using "mlm" (Masked Language Model) parameter in the training process. I though that the …
-
-
### 🚀 The feature, motivation and pitch
there's a new DP shard strategy which is more flexible and general, see more detail at https://arxiv.org/abs/2311.00257 AMSP: Reducing Communication Overhead o…
-
![image](https://github.com/paperswithlove/papers-we-read/assets/12858045/20087322-d388-45db-b0ed-2daab0ea5baf)
[https://arxiv.org/abs/2403.09611](https://arxiv.org/abs/2403.09611)
- 아니 애플에서 MLL…
-
Dear authors,
Thanks for your work! I am interested in applying it in my study. I wonder could you provide the fine-tuned WizardCoder model file, which could be ready-for-use. Or could you pleas…
-
Hello,
I have a question regarding GPU memory consumption during inference.
Before finetuning a model with QLora, the torchtune.LoRALinear modules will convert the original LLM weights to nf4, a…
-
In the demos I’ve seen of Leon AI, it appeared rather slow. I have no idea if this was a limitation of the hardware or there were inefficiencies that might be improved upon. [GPT4All](https://github.c…
-
We discussed here: https://github.com/kubeflow/website/pull/3718#issuecomment-2096619898 that [our LLM Trainer](https://github.com/kubeflow/training-operator/blob/bb8bba00ff0b48de922c523b0d3051f8b2d4e…