-
accelerate launch --config_file "C:/Users/1/.cache/huggingface/accelerate/default_config.yaml" train.py --model_name_or_path Shitao/OmniGen-v1 --batch_size_per_device 1 --condition_dropout_prob 0.01 -…
-
Hi~I am trying to reproduce the results of RepLLaMA. I have an a800 GPU. If I start training RepLLaMA from scratch with your code, it may take 80 hours? I want to know if this is normal? If possible, …
-
- We have executed the LLaMA model (downloaded from HuggingFace[https://huggingface.co/meta-llama/Llama-2-7b-chat-hf]) using the NNTrainer and obtained the following output by following these steps…
-
Are there any plans to support bf16 training in deepspeed in the near future?
If not - could someone guide me toward what I would need to change in order to implement it? It seems like a fair few thi…
-
When creating`model_2` and trying to load the weights by
```python
model_2.load_weights(checkpoint_path)
```
I'm getting the following error:
```
-----------------------------------------…
-
### Search before asking
- [X] I had searched in the [issues](https://github.com/apache/doris/issues?q=is%3Aissue) and found no similar issues.
### Version
1.2.6
### What's Wrong?
There are thre…
-
### Checklist
- [X] I've read the [contribution guidelines](https://github.com/autowarefoundation/autoware/blob/main/CONTRIBUTING.md).
- [X] I've searched other issues and no duplicate issues were…
-
did you successfully train on the 10M web videos?
I'm looking at exploring this suggestion for 4 lies of code change.
https://github.com/guoyww/AnimateDiff/issues/308
-
Thanks for your excellent work. Would you mind me asking what is the pretraining acc on imagenet2012 that then used for finetuning?
-
- [ ] [LMOps/README.md at main · microsoft/LMOps](https://github.com/microsoft/LMOps/blob/main/README.md?plain=1)
# LMOps/README.md at main · microsoft/LMOps
## LMOps
LMOps is a research initiati…