-
When finetuning the model I will have to create a dataset of some episodes.
Do you have any resources on how you recorded/created a RLDS dataset? It seems to be somewhat niche with little documentat…
-
Thanks for your impressive work. But the dataset given in the Baidu disk doesn't fit the finetuning process. If I want to reprodce the results in your paper, should I download the official SUN RGB-D d…
-
lightning-cli getinfo
```
{
"id": "035aef5661e1a6e370db60dc0455796800afd5b51fbc12a0a8b34836b15f5d7ef6",
"alias": "TWronald✅",
"color": "15c315",
"num_peers": 9,
"num_pending_chann…
-
Good Job! Do you have a plan to support LoRA or other PEFT?
-
🦥 Unsloth: Will patch your computer to enable 2x faster free finetuning.
==((====))== Unsloth: Fast Llama patching release 2024.6
\\ /| GPU: NVIDIA A100 80GB PCIe MIG 7g.80gb. Max memory: 7…
-
That sounds massively interesting, and while we try to run inference and read the paper, should we expect the release of the finetuning code?
-
I was trying to finetune a model as mentioned in the docs, but after training, when I am trying to load the model, I get the following error:
```
[Error(s) in loading state_dict for SubwordBert:
s…
-
Hi,
Thanks for the great work. Is it possible to release the data (and codes) used to fine-tune the language model.
-
Hello,
I am trying to finetune the tapas_wtq_wikisql_sqa_masklm_medium_reset.
Just to see, if it works in general, I wanted to finetune it on the same data it's already trained on, WTQ. Creating …
-
First of all, great job on the project! I was wondering if there were any implemented functionality to finetune this model? Basically introducing another dataset to train on top of the already trained…