Open breezedeus opened 1 year ago
Hi, I have not tried or verified the full model finetuning using int8 training. Int8/Int4 are mainly designed for quantized LoRA training. Please check out the instructions here.
Hi, I have not tried or verified the full model finetuning using int8 training. Int8/Int4 are mainly designed for quantized LoRA training. Please check out the instructions here.
@haotian-liu Thanks for your suggestion. I know there were problems with the previous training, so I changed to lora+8-bits training as you suggested. Below is my instruct tuning loss curve:
The results generated from the model are not very satisfactory. For instance, I want the model to describe the llava logo image:
python llava/eval/run_llava.py \
--model-path ./checkpoints/llava-vicuna-v1-3-7b-finetune-lora \
--model-base lmsys/vicuna-7b-v1.3 \
--load-8bit \
--image-file images/llava_logo.png \
--query "describe the image"
The following are generated responses:
The image is a close-up of a person's face, with a blurred background. The person's eyes are open, and their facial expression appears to be one of concentration or focus. The
blurred background adds a sense of depth and focus to the image, making the person's face the primary point of interest. The overall effect is a visually striking and engaging
portrait that captures the viewer's attention.
Would you mind providing your Lora instruct tuning logs? Thanks much.
Question
Many thanks to the authors for this very good work.
I'm trying to run the instruct tuning part on one 3090 GPU, using the int-8 mode:
The above command can finish the training and save the model successfully. The following is the loss curve:
But when I use the following command to make predictions using the trained model, the result is very bad:
It's the result:
Does anyone know why? Is it because I'm using the model incorrectly? Or is this model less effective in int-8 mode?
Thanks much.