-
Hello,
I'm currently training LLaMA PRO. Initially, I expanded the model from 32 layers to 40 layers and proceeded to train only the newly added 8 layers (every fifth layer). Therefore, I froze 32 …
-
200N updated
3 months ago
-
Hi @deJQK,
Seems like you are using apache arrow for distributed training.
Can you explain more on how to configurate environment for pyarrow?
I cannot start training as i always get `FileNotFou…
-
(nanodet) simon@Simon:~/nanodet$ python tools/train.py config/nanodet-m-416.yml
[root][07-16 11:17:37]INFO:Using Tensorboard, logs will be saved in workspace/nanodet_m_416/logs
[root][07-16 11:17:37…
-
I have a question about distributed training, how to run the idm_main.py file on my single graphics card window computer.My question is that "RuntimeError: Default process group has not been initializ…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
transformers/training_args.py", line 1712, in __setattr__
raise FrozenInstanceError(f"ca…
-
并行化模型的时候可以加上一步操作:
```python
# Convert BatchNorm to SyncBatchNorm.
net = nn.SyncBatchNorm.convert_sync_batchnorm(net)
```
确保batch norm 在所有process上sync了。
参考:
https://theaisummer.com/distribute…
-
Hello authors,
Thanks so much for sharing these codes.
The codes are very useful to fine-tune SAM for downstream works : )
I reduced datasize, adapted the codes and run them in **Google Colab w…
-
Hello,
I noticed a deviation from the Griffin paper in your code.
The Griffin paper states in the second part of chapter 2.4:
> We initialize Λ such that a^c is uniformly distributed between 0.…
-
你好,请教一下,训练的时候,出现如下问题:
`cd Chinese-CLIP/
bash run_scripts/muge_finetune_vit-b-16_rbt-base.sh ${DATAPATH}`
出现下面的问题:
`root@clip-test-d9cd48656-q2zbl:~/workspace/clip/Chinese-CLIP# bash run_scripts/…