Closed Tengfei000 closed 1 week ago
UserWarning: None of the inputs have requires_grad=True. Gradients will be None
--version llama
It works!Thanks
Thanks!When I use the “fineturn_lora.sh”, I still can not get the right results. Here is my setting.
fineturn_lora.sh
MODEL_TYPE=llama3-8b
PRETRAIN_DIR=bunny-$MODEL_TYPE-pretrain OUTPUT_DIR=bunny-lora-$MODEL_TYPE
mkdir -p ./checkpoints-$MODEL_TYPE/$OUTPUT_DIR
deepspeed bunny/train/train.py \ --lora_enable True --lora_r 128 --lora_alpha 256 --mm_projector_lr 2e-5 \ --deepspeed ./script/deepspeed/zero3.json \ --model_name_or_path ./path/to/llama3-8B \ --model_type $MODEL_TYPE \ --version llama \ --data_path ./finetune/bunny_695k.json \ --image_folder ./finetune/images \ --vision_tower ./path/to/siglip-so400m-patch14-384 \ --pretrain_mm_mlp_adapter ./checkpoints-pretrain/$PRETRAIN_DIR/mm_projector.bin \ --mm_projector_type mlp2x_gelu \ --image_aspect_ratio pad \ --group_by_modality_length False \ --bf16 True \ --output_dir ./checkpoints-$MODEL_TYPE/$OUTPUT_DIR \ --num_train_epochs 1 \ --per_device_train_batch_size 8 \ --per_device_eval_batch_size 4 \ --gradient_accumulation_steps 2 \ --evaluation_strategy "no" \ --save_strategy "steps" \ --save_steps 500 \ --save_total_limit 1 \ --learning_rate 2e-4 \ --weight_decay 0. \ --warmup_ratio 0.03 \ --lr_scheduler_type "cosine" \ --logging_steps 1 \ --tf32 True \ --model_max_length 2048 \ --gradient_checkpointing True \ --dataloader_num_workers 4 \ --lazy_preprocess True \ --report_to none | tee 2>&1 ./checkpoints-$MODEL_TYPE/$OUTPUT_DIR/log.txt
Close the issue for now if there's no further discussions. Feel free to reopen it if there's any other questions.
I employed siglip and llama3 for training. The default “finetune_full.sh” and data are used, but my loss is 0.