-
I have the following error when finetune the DocOwl1.5-Omni. It always raises error when index is 10. Please help!!!
```
File "/opt/conda/envs/mplug_owl2/lib/python3.10/site-packages/deepspeed/run…
-
Hi @gabrieltseng, I've read your paper and find it a really interesting work!
Thanks a lot for sharing your code as well!
I'm trying to adapt your downstream task [notebook](https://github.com/na…
-
Good Job! Do you have a plan to support LoRA or other PEFT?
-
Hello everyone, below is my code for fine-tuning XTTS for a new language. It works well in my case with over 100 hours of audio.
https://github.com/nguyenhoanganh2002/XTTSv2-Finetuning-for-New-Lang…
-
Hi there,
First of all thanks for such an awesome work, I tried it for my custom usecase, it gave me awesome results.
But only 2 classes are missing, I also tried using different custom labels, as t…
-
Hi, thanks for sharing this wonderful work. Since you use the multi-frame multi-view inputs during pretraining stage, I want to know whether did you still use the temporal multi-frame inputs during fi…
-
在finetune代码中,部分也加入了loss的计算,想请教下这样相比conditioning language modeling loss有什么特别的好处吗?
-
I was trying to finetune a model as mentioned in the docs, but after training, when I am trying to load the model, I get the following error:
```
[Error(s) in loading state_dict for SubwordBert:
s…
-
Hi,
Thanks for the great work. Is it possible to release the data (and codes) used to fine-tune the language model.
-
Hello,
I am trying to finetune the tapas_wtq_wikisql_sqa_masklm_medium_reset.
Just to see, if it works in general, I wanted to finetune it on the same data it's already trained on, WTQ. Creating …