-
### Describe the issue
**Issue:**
I ran into tokenization mismatch errors when I tried to fine-tune from Llama-3.1. I pre-trained a new MLP adapter for Llama-3.1 and that seems to work, but the fine…
-
Hey guys,
I'm pretty new here just trying to figure all this out.
Finally managed to get my first finetuning running. But I'm kinda confused.
I'm using the thomas - medium model (german) for fine…
-
When finetuning the model I will have to create a dataset of some episodes.
Do you have any resources on how you recorded/created a RLDS dataset? It seems to be somewhat niche with little documentat…
-
# I am planning to perform a two-stage fine-tuning process and need some guidance on how to proceed.
## First Stage
1. Load Base Model: I start by loading the base model, qwen1.5 32B.
2. Appl…
-
During the use of LoRA fine-tuning, everything was normal, but the following issue arose during full-scale fine-tuning.
I use the following script for full fine-tuning :
```shell
#!/bin/bash
N…
-
Hello!
I am trying to finetune either the vit_s or vit_b models to my dataset. I have tried training only the dino head, both the dino and ibot heads, and keeping the whole backbone frozen or unfr…
-
Thanks for your impressive work. But the dataset given in the Baidu disk doesn't fit the finetuning process. If I want to reprodce the results in your paper, should I download the official SUN RGB-D d…
-
lightning-cli getinfo
```
{
"id": "035aef5661e1a6e370db60dc0455796800afd5b51fbc12a0a8b34836b15f5d7ef6",
"alias": "TWronald✅",
"color": "15c315",
"num_peers": 9,
"num_pending_chann…
-
That sounds massively interesting, and while we try to run inference and read the paper, should we expect the release of the finetuning code?
-
Hi, thanks for sharing this wonderful work. Since you use the multi-frame multi-view inputs during pretraining stage, I want to know whether did you still use the temporal multi-frame inputs during fi…