-
PyTorch has several backends for [distributed computing](https://pytorch.org/docs/stable/distributed.html). One of them is NVidia NCCL where main benefit is distributed computing for most Devices on G…
-
When I generate a variable font, the `public.fontInfo['familyName']` doesn't seem to be considered during compilation.
I see there has been some work done on the fontTools side. Is there still some w…
-
**Is your feature request related to a problem? Please describe.**
I would like to be able to use multiple GPUs to generate multiple images at a time when using the `diffusers` backend.
**Desc…
-
Hello!Thank you so much for your work
I would like to ask is there any effect on removing distributed training from model training.
Thank you!
-
## Purpose
We want to implement distributed matrix multiplication to enable parallel online regridding in the coupler.
## Cost/Benefits/Risks
* Costs:
* Developer time
* Risks:
* Incorrect regrid…
-
### 🐛 Describe the bug
Hi,
When I launched multi-process training (8x A100) using `torchvision.datasets.ImageNet()` with fresh-prepared `root` (i.e. containing only `ILSVRC2012_devkit_t12.tar.gz…
EIFY updated
17 hours ago
-
Hi !
I'm trying to work on Windows with a cross-platform toolchain so I'm using Ninja + Clang for my project. Since Clang for MSCV is ABI compatible with MSVC code, everything works fine with the …
-
I'm serving our IIFE builds as fireproof.iife.js and connect.iife.js respecitvely (both are distributed as an index.global.js in their packages).
As currently shipped they contain the following ref…
-
[rank1]:[W CUDAGraph.cpp:145] Warning: Waiting for pending NCCL work to finish before starting graph capture. (function operator())
[rank0]:[W CUDAGraph.cpp:145] Warning: Waiting for pending NCCL wor…
-
### Describe the bug
Thank you for your amazing work. It seems like models are not saved or loaded properly after finetuning train_custom_diffusion.py in a new dataset. Generated validation images ar…