-
### Describe the bug
I was trying to run the lora training script of PixArt-alpha: https://github.com/PixArt-alpha/PixArt-alpha/blob/master/train_scripts/train_pixart_lora_hf.py but got Runtime Error…
-
I use custom data to train DINO, the model seems collapsed after a few steps, the feature seems to be uniform. I use larger teacher temputure to enhance "sharping", but the model collapsed after all.…
-
Impressive and insightful work, hooray to the authors! Recently I read your paper, but I'm comfused about the following parts.
1. In the abstract, you discuss how memory-reduction approaches like LoR…
-
## Description
Relatively minor, but explicitly omitting `allow-same-origin` from the help widget iframe `sandbox` attribute in packages/help-extension breaks search pages on many reference documen…
-
-
hi lin, i managed to write a finetuning script, could you help me check it? i also got confused about some details, listed below(also marked with NOTE in code comments), could you illustrate somehow? …
-
### System Info
When I used P-tuning V2 to fine-tune GLM, the loss reduction was very noticeable, but in the actual inference, I made a lot of noise. I use the training data again for inference.
Exa…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### Reproduction
参考https://github.com/hiyouga/LLaMA-Factory/wiki/Performance-comparison,使用 [llama 8B](https://hf-mirror.c…
-
### Reminder
- [X] I have read the README and searched the existing issues.
### System Info
bin C:\Users\luoxiaojie\.conda\envs\pytorch212-lxj\lib\site-packages\bitsandbytes\libbitsandbytes_cuda121…
-
I am playing with the MMA-hard model to replicate WMT15 DE-EN experiments reported in the paper and my question is about preprocessing and postprocessing data. The paper says that:
> For each data…