-
### Prerequisites
- [X] I have read the [documentation](https://hf.co/docs/autotrain).
- [X] I have checked other issues for similar problems.
### Backend
Hugging Face Space/Endpoints
### Interfac…
-
I have the following error when finetune the DocOwl1.5-Omni. It always raises error when index is 10. Please help!!!
```
File "/opt/conda/envs/mplug_owl2/lib/python3.10/site-packages/deepspeed/run…
-
Good Job! Do you have a plan to support LoRA or other PEFT?
-
I was trying to finetune a model as mentioned in the docs, but after training, when I am trying to load the model, I get the following error:
```
[Error(s) in loading state_dict for SubwordBert:
s…
-
在finetune代码中,部分也加入了loss的计算,想请教下这样相比conditioning language modeling loss有什么特别的好处吗?
-
Hi,
Thanks for the great work. Is it possible to release the data (and codes) used to fine-tune the language model.
-
Hello,
I am trying to finetune the tapas_wtq_wikisql_sqa_masklm_medium_reset.
Just to see, if it works in general, I wanted to finetune it on the same data it's already trained on, WTQ. Creating …
-
First of all, great job on the project! I was wondering if there were any implemented functionality to finetune this model? Basically introducing another dataset to train on top of the already trained…
-
Should finetune all TDBN or just CRBM+LogReg?
At the moment (13 avril 2016), I try to put together all my models.
In other words, until now I train RBMs and use them to generate a new dataset (train,…
-
My current guess is that any fastchat-based code would work: create the model, apply all the patching you need, and then train?
The only question I have is about the dataset: does it have to contain…