-
can you please modify modeling_xgml.py . In such way, we can you your code with huggingface. Thank you
-
I ran 'train_minillm.py' successfully under the guidance of the README.md file. However, due to some uncontrollable factors, the GPU will interrupt approximately every 6-8 hours. At this time, the loc…
-
I tried to use a simlar dataset alpaca-zh to sft the llama-7b on 16 x 32G v100 gpus. gpu_per_node=8 ,node_num=2.
The script I use is `scripts/llama/sft/sft_7B.sh`.
But the training loss did not …
-
**What steps does it take to reproduce the issue?**
* When does this issue occur?
While trying to work around the ListSets command not working in the GUI (see #8289 )...
I tried to first create …
-
I was reading [this arxiv article](https://arxiv.org/pdf/2304.04487.pdf) which points to the current repo (https://github.com/microsoft/unilm).
The article describes a method to speed-up inference …
-
Hey, I'm following the guide in the [readme](https://github.com/microsoft/LMOps/blob/b95509cb6e5a6c8bedbea936cd353735521a91d8/structured_prompting/fairseq-version/README.md) to run the manyshots struc…
-
I try to distill gpt2-1.5B -> gpt2-120M
As I use 4 A100, so I change the GPUS_PER_NODE to ${3-4}
Batch size remains the same
-
```
m1:/opt/ocfs2-test/bin # bash run.sh
Mon Sep 7 10:20:39 CST 2015
xattr-test.................................................. Passed.
Runtime 42255 seconds.
Mon Sep 7 22:04:54 CST 2015
inli…