-
when use 34B LLM model, I can not put all the model to single gpu. so I use device_map='auto' to put several part to multi gpus.
but I found inference time cost too much, so how to solve this problem…
-
Dear All,
I would like to run ALIGNN on multi GPUs. When I checked the code I could not find any option.
Is there any method to run ALIGNN on multi GPUs such as using PyTorch Lightning or DDP fu…
-
### 🐛 Describe the bug
I met an error when I use torchrun for 4 GPUs training and 'nccl' backend (It runs perfect when I use 'gloo'). The environment is python3.9+pytorch2.3.0+cuda12.1.We tried to …
-
Hi,
The multi-gpu setting does not work.
return forward_call(*input, **kwargs)
File "/home/hossein/projects/Ladder-Side-Tuning-main/seq2seq/third_party/models/t5/modeling_side_t5.py", line …
-
great work guys, I was just curious if there is going to be support for multiple gpu's in the future.
Thanks
-
@AddyLaddy
Does FM support partitioning one compute node for multi-tenant? For example, I have 4 GPUs in one compute node, but I want to partition them into 2 nvlink groups, however they can't commun…
-
**Describe the bug**
When running auto3d to generate 3D structures with multiple GPUs (e.g., using --gpu_idx 0,1), I encounter an OSError related to an invalid file, arated.sdf. This error does not o…
-
Hi,
Thanks for the woderful job.
I encountered a error caused by distributed training, maybe? I ran the code on multi-gpus and got the error below:
`RuntimeError: Expected to have finished reductio…
-
### Feature request
I want to use accelerate for multi-machine and multi-gpus inference. Since using trainer.predict does not return the original inference data, only the inference results, I am not …
-
Could u update code for multi-GPUs training?
I've tried change it for multi-GPUs training, but got some bugs
I've changed `train.py` as follows:
```python
if torch.cuda.device_count() > 1:
…