-
When I finetune the animate-anything, I found that the gradient of unfrozen layer of Unet (eg. conv_in) is None. And I print the requires_grad of conv_in, the result is True. It means that the fine-tu…
-
That sounds massively interesting, and while we try to run inference and read the paper, should we expect the release of the finetuning code?
-
# Title of the Talk: No Code SLM Finetuning with MonsterAPI
## Abstract of the Talk:
Dive into the world of no-code large language model (LLM) finetuning in this informative talk presented by Mons…
-
While executing the file in folder `Olive/examples/llama2` i got the error of
TypeError: LlamaForCausalLM.forward() got an unexpected keyword argument 'past_key_values.0.key'
while executing :
`py…
-
在finetune代码中,部分也加入了loss的计算,想请教下这样相比conditioning language modeling loss有什么特别的好处吗?
-
Hello Xintao,
We found that the direct inferencing based on GFPGAN v1.4 performs pretty well on our own datasets, whilst GFPGAN v1 inferencing is not high-quality.
However, when we tried to fin…
-
My current guess is that any fastchat-based code would work: create the model, apply all the patching you need, and then train?
The only question I have is about the dataset: does it have to contain…
-
I was trying to finetune a model as mentioned in the docs, but after training, when I am trying to load the model, I get the following error:
```
[Error(s) in loading state_dict for SubwordBert:
s…
-
Hi,
Thanks for the great work. Is it possible to release the data (and codes) used to fine-tune the language model.
-
Hello,
I am trying to finetune the tapas_wtq_wikisql_sqa_masklm_medium_reset.
Just to see, if it works in general, I wanted to finetune it on the same data it's already trained on, WTQ. Creating …