-
Traceback (most recent call last):
File "/data/ChatGLM-Tuning/finetune.py", line 117, in
main()
File "/data/ChatGLM-Tuning/finetune.py", line 110, in main
trainer.train()
File "/ro…
-
Hello. Thank you for sharing such a great work. I am trying to run samples in inference.py. The instruction-tuned worked perfectly. However, the in-context working example for pretrained model did not…
-
### 🐛 Bug
Today when attempting to upload a LoRA-trained Llama 3.1 70B model (first time I've trained Llama 3.1), I hit the following during the eLoRA merge. Note I used the `cpu_shard` method to u…
-
### 是否已有关于该错误的issue或讨论? | Is there an existing issue / discussion for this?
- [X] 我已经搜索过已有的issues和讨论 | I have searched the existing issues / discussions
### 该问题是否在FAQ中有解答? | Is there an existing ans…
-
We are hosting a model in SageMaker and today we observed the following error in our logs when the model was being relaunched in the instance:
```
ImportError: cannot import name 'Conversation' fr…
-
For Mistral-7B-v0.1 20% compression ratio, PPL after pruning: {'wikitext2': 245.2660781818917}, the pruned ppl for Llama-3-8B is also bad like Mistral-7B.
Could you tell me how to reproduce the resul…
-
### Describe the bug
**Problem:**
When the "officially provided" example (see https://github.com/huggingface/diffusers/blob/a785992c1d6fcb1ff66f8a0d68d09a0a81b909e8/src/diffusers/pipelines/ledits_pp…
-
Dear Author,
Thanks for your great projects.
I was trying to evaluate the model without Tuning and with Tuning. I wondered if we can evaluate the model with the original model.
Also, if I want to…
-
In my training script, I set the **per_device_train_batch_size = 4** in the TrainingArguments.
But the **train_batch_size** in the **trainer_state.json** of each checkpoint is **2**.
When I tried …
-
# URL
- https://arxiv.org/abs/2305.14314
# Affiliations
- Tim Dettmers, N/A
- Artidoro Pagnoni, N/A
- Ari Holtzman, N/A
- Luke Zettlemoyer, N/A
# Abstract
- We present QLoRA, an efficient fi…