-
Hi, thanks for your work, I recently wanted to try multi-GPU training, but I realized that its default is to use DataParalle instead of DDP, can you tell me where I can switch to DDP mode?
-
When I run the tutorial [Example: Deep structural causal model counterfactuals](https://basisresearch.github.io/chirho/deepscm.html), the following cell:
```Python
adam_params = {"lr": 1e-3}
batch_…
-
I add my username and API token to configuration file:
...
neptune_username: yhpmaster
neptune_api_key: ******
...
But when I train, the following error occurs:
...
[root@b62ad5cd2324 home]# dl…
yihp updated
4 months ago
-
This was a known issue last year with PyTorch version issues, but it was never written down permanently (only in Slack messages that have since been automatically deleted)
```
`(amptorch) [jparas7@l…
-
### Description
When making a decision against the Player, the AI should consider:
* ~~**Has** the enemy Pokémon undergone the Terastal phenomenon? If so, what is its new type?~~
* ~~**Could** …
-
### Bug description
I am trying to train a GAN model on multiple GPUs using DDP. I followed the tutorial at https://lightning.ai/docs/pytorch/stable/notebooks/lightning_examples/basic-gan.html, chang…
-
```
{
"name": "CompilationError",
"message": "at 53:4:
loss_ptr += row_idx
logsumexp_ptr += row_idx * N_CHUNKS + chunk_idx
labels_ptr += row_idx
col_offsets = chun…
-
Observations:
- Using fused optim (e.g. Adam) + `ipex.optimize(..., level='01', dtype=torch.bfloat16)` **no error**
- non fused optimizer (e.g. AdamW) + `ipex optimize(..., level='01', dtype=torch.f…
-
╭─────────────────────────────── Traceback (most recent call last) ────────────────────────────────╮
│ /home/mdisk2/tanjunwen/gitprj/mChatGLM_mutli_gpu_tuning/finetune_ptuning.py:34 in │
│ …
-
hello~I will be appreciate if you can give any advice!
my accelerate,peft,deepspeed,transformer are all the latest version. "device_map" is set to auto
CUDA_VISIBLE_DEVICES="0,5" torchrun --nnod…