-
-
### Search before asking
- [X] I have searched the YOLOv8 [issues](https://github.com/ultralytics/ultralytics/issues) and [discussions](https://github.com/ultralytics/ultralytics/discussions) and f…
-
### 🚀 The feature, motivation and pitch
DDP bucket will always in GPU HBM,which size is same as the sum of module all weight gradients' size.
In fwd stage and optimizer stage, this memory is wast…
-
### 🚀 The feature, motivation and pitch
**Background**
DistributedDataParallel (DDP) uses `Reducer` to bucket and issue `allreduce` calls. The main entry point of `Reducer` is through the gradient …
-
Dear authors,
Thank you for sharing the code. In the paper, you have mentioned that the Cal-GAN is trained using 4 GPUs and in the repo you kindly provided multigpu training instructions. However, …
-
"When I used this command, I did not use multiple GPUs, what is the reason?
`python train.py --data example.yaml --adam --norm CT --epochs 1000 --patience 200 --device 4,5,6,7`
-
Hi, first of all, thank you for this cool work! It's impressive and I appreciate the effort you've put into it.
I have a question about using Flora with DDP. Have you tried Flora to train 7B with DDP…
-
What's the "best-practice" for configuring CachedMultipleNegativesRankingLoss when used for DDP. Say for example I have 3000 unique `positive` labels in my dataset, and I'm training using DDP on a sin…
-
Hi,
I'm having some issues with the training for blendedmvs using DDP mode.
`Traceback (most recent call last):
File "train.py", line 265, in
mp.spawn(main, nprocs=args.world_size, args=(…
-
[h264 @ 0x16543c00] Missing reference picture, default is 65562
[h264 @ 0x16543c00] mmco: unref short failure
[h264 @ 0x16543c00] mmco: unref short failure
[h264 @ 0x16543c00] Missing reference pic…