-
🦥 Unsloth: Will patch your computer to enable 2x faster free finetuning.
==((====))== Unsloth: Fast Llama patching release 2024.6
\\ /| GPU: NVIDIA A100 80GB PCIe MIG 7g.80gb. Max memory: 7…
-
Good Job! Do you have a plan to support LoRA or other PEFT?
-
# URL
- https://arxiv.org/abs/2310.05914
# Affiliations
- Neel Jain, N/A
- Ping-yeh Chiang, N/A
- Yuxin Wen, N/A
- John Kirchenbauer, N/A
- Hong-Min Chu, N/A
- Gowthami Somepalli, N/A
- …
-
I was trying to finetune a model as mentioned in the docs, but after training, when I am trying to load the model, I get the following error:
```
[Error(s) in loading state_dict for SubwordBert:
s…
-
Hi,
Thanks for the great work. Is it possible to release the data (and codes) used to fine-tune the language model.
-
Hello,
I am trying to finetune the tapas_wtq_wikisql_sqa_masklm_medium_reset.
Just to see, if it works in general, I wanted to finetune it on the same data it's already trained on, WTQ. Creating …
-
First of all, great job on the project! I was wondering if there were any implemented functionality to finetune this model? Basically introducing another dataset to train on top of the already trained…
-
Should finetune all TDBN or just CRBM+LogReg?
At the moment (13 avril 2016), I try to put together all my models.
In other words, until now I train RBMs and use them to generate a new dataset (train,…
-
While executing the file in folder `Olive/examples/llama2` i got the error of
TypeError: LlamaForCausalLM.forward() got an unexpected keyword argument 'past_key_values.0.key'
while executing :
`py…
-
Hello Xintao,
We found that the direct inferencing based on GFPGAN v1.4 performs pretty well on our own datasets, whilst GFPGAN v1 inferencing is not high-quality.
However, when we tried to fin…