-
I have a 2 GPU system, a 3060 (12gb VRAM) and a 3070ti (8GB). I've read torch supports paralellism that can split large models into both GPUs, it'd be great to have something like that to run big mode…
-
When I use multi-GPU training, I encounter the following problem:
subprocess.CalledProcessError: Command '['/home/a/anaconda3/envs/mambayolo/bin/python', '-m', 'torch.distributed.run', '--nproc_per…
-
i wonder how to put it on a machine with multi-GPU to accelerate its training?
-
hello, I encountered some problems while using this code for multi-gpu training.
first I tried to run it with
"python3 train_dafnet.py --model_name "llama-2-7b" --device 0 --extra_device 1 2 3"
an…
-
I ran mem_spd_test.py and got the following error:
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:0 and cuda:1!
I did not make any changes except …
-
Hello, have you tried training on multi-gpus? One error exists:
![image](https://github.com/user-attachments/assets/ef02add7-83b8-4d65-884a-13eb9dadc30f)
-
### System Info
- `transformers` version: 4.45.0.dev0
- Platform: Linux-4.18.0-477.10.1.el8_8.x86_64-x86_64-with-glibc2.28
- Python version: 3.11.5
- Huggingface_hub version: 0.24.0
- Safetenso…
-
🚀 The feature, motivation and pitch
# RFC: Multi-Gpu Python Frontend API
This RFC compares and contrasts some ideas for exposing multi-gpu support in the python frontend.
1. The current `multigpu_sc…
-
### Feature Idea
Allow memory to split across GPUs. With the arrival of Flux, even 24gb cards are maxed out and models have to be swapped in and out in the image creation process, which is slow. If y…
-
Is there something like multip-GPU support for this? Would be awesome due to long processing times.