-
I am using LLama-Factory to train long text DPO, but enabling unsloth is not supported with the latest version of the trl library. The newest trl update includes many useful new alignment algorithms. …
-
Hi,
when I try to train model with multi GPUs(CUDA_VISIBLE_DEVICES=0,1,2 python3 train.py …), I got an error like 'core dumped',while train with single GPU is fine
best
ghost updated
3 years ago
-
Hello! Thanks for sharing your excellent work! I tried the training code and have the following two problems.
1) I have tried training with 2 GPUs using the command in run.sh. But the training stop…
-
I don't find the parameter for multi-gpus training
-
**Is your feature request related to a problem? Please describe.**
Cannot use multiple GPU
**Describe the solution you'd like**
Allow for multiple GPU training
reported by @biancazadrozny
-
Nice work! And I wonder how to train on multi gpus instead of single gpu? Can you provide the code plz? Thanks a lot!
-
Hi,
Thanks for your great work MAML++.
I'm having a problem with multi-GPU. When I input
``` bash
bash experiment_scripts.sh 0 1 2
```
to run the training, it seems that only the first GPU ID…
Apscg updated
11 months ago
-
Nowadays human baselines run in a container on a VP machine, then the container is stopped, and eventually it's later restarted to score it and otherwise look at what was done in there. With multi-nod…
-
Hello, quark!
Thx for your great work. When I tried to run your train_search job with multi-gpus, the Variable of alphas_normal and alphas_reduce causes errors.
The errors are shown as followi…
-
in nougat's options, we don't need to allocate GPUs. we have multiple GPUs in the server, can nougat use multi GPUs in one task?
if nougat can only use one GPU in one task, can we run serveral tasks …