-
**Describe the bug**
What the bug is, and how to reproduce, better with screenshots(描述bug以及复现过程,最好有截图)
The channel dimension is ambiguous. Got image shape (3, 672, 3). Assuming channels are the firs…
-
I find Microsoft's Phi 3.5 vision instruct performs much better than Florence 2. Since it's an instruct model, it also has the benefit of taking text instruction as input to help describing the images…
-
Hi,
Thank you for awesome library!
I am using `litgpt version 0.4.11`
Currently I am using `Phi-3.5-mini-instruct` to finetune using lora. Even though I set `--train.max_seq_length 10000` I st…
-
![image](https://github.com/user-attachments/assets/e39441e9-6de6-483e-92a2-612dfb02811d)
使用的命令是:
bash finetune.sh -m models/models--Qwen--Qwen2-7B-Instruct/snapshots/41c66b0be1c3081f13defc6bdf946c2…
-
### System Info
I am finetuning Llama3-8b-Instruct model. Here is the Jupyter Notebook of the steps, i followed to perform the finetuning:
https://gitlab.com/keerti4p/llama3-8b-instruct-finetune/-…
-
Really appreciate for the youtube video for finetune with Mac M1,
And I can run the finetune successfully on my Mac M1.
```
python scripts/lora.py --model mlx-community/Mistral-7B-Instruct-v0.2-4b…
-
Hello Unsloth Team,
I am trying to finetune the **dwb2023/phi-3-vision-128k-instruct-quantized** model using Unsloth, but I encountered a NotImplementedError. The error message indicates that this …
-
I followed the current Google Colab notebook for finetuning Llama 3.1 8B Instruct, which includes the use of `train_on_responses_only`. `train_on_responses_only` adds a `labels` column to `trainer.tr…
-
Three models
* The pre-trained model (model A) is used as the base model.
* We finetuned these base models by using our own dataset (Get model B). // this is used as enhanced models
* Instruct tune…
-
Please correct me if I'm wrong, but it looks like the current examples for lora training all build a loss function around completion, which lines up with the lora example of using only the 'text' fiel…