-
Is there any chance that we can have a full description on how to fine-tune pre-trained models (for example in machine translation)? I've managed to continue training on a much smaller dataset (by usi…
y3nk0 updated
4 years ago
-
### Question
The instruction tuning data, such as llava_instruct_158k.json, can not be downloaded. Can you update the dataset?
-
Hello, thanks a lot for sharing your training code!
In the training script (https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/scripts/train/finetune_onevision.sh), is it possible to use the trained L…
-
![image](https://github.com/user-attachments/assets/c1975862-540e-4b3d-9304-c3b7a0f31020)
Description: During training on the Expository-Prose-V1 dataset using the provided parameters, I observed…
-
Dear Author,
I have fine-tuned the ABR model according to the instructions using the provided hyperparameters. However, the results I obtained are noticeably different from the results reported in …
-
Hello I'm in the process of fine-tuning a Large Language Model (LLM) for an NGO and I need to construct an instruction dataset from .pdf and .docx documents containing information in text.
The obje…
-
I have placed your instruction data into the dataset for the first stage of fine-tuning, but at 27% progress, the loss seems to have become zero, and the loss has been close to zero throughout the pro…
-
-
Hello, first of all thank you for your amazing work!
I wonder if you have any plans for open-sourcing the Emotion Instruction Tuning data you created for instruct tuning the model mentioned in the…
-
Hi, please provide additional Instruction about How to fine-tuning QWEN model with BigDL speed up.