hiyouga / LLaMA-Factory

Unify Efficient Fine-Tuning of 100+ LLMs
Apache License 2.0
25.26k stars 3.13k forks source link

[Help] Yi-34B + Simpo + Full + Novel writing task, fine-tuning results are unsatisfactory! #4563

Closed onlyfish79 closed 2 days ago

onlyfish79 commented 2 days ago

Reminder

System Info

Reproduction

torchrun --nproc_per_node=8 --master_port=20001 src/train.py \
--stage dpo \
--pref_loss simpo \
--pref_beta 2.0 \
--simpo_gamma 1.0 \
--do_train \
--model_name_or_path /data/basemodel/Yi-34B \
--dataset novel \
--val_size 0.05 \
--dataset_dir /data/train_data \
--overwrite_cache \
--template yi \
--finetuning_type full \
--output_dir /data/dpo_model/yi34b_simpo_full \
--overwrite_output_dir \
--preprocessing_num_workers 128 \
--per_device_train_batch_size 4 \
--per_device_eval_batch_size 8 \
--gradient_accumulation_steps 4 \
--eval_strategy steps \
--eval_steps 10 \
--save_strategy steps \
--save_total_limit 1 \
--lr_scheduler_type cosine \
--logging_steps 1 \
--save_steps 230 \
--learning_rate 1e-5 \
--num_train_epochs 1.0 \
--max_grad_norm 0.5 \
--weight_decay 0 \
--warmup_ratio 0.03 \
--cutoff_len 2048 \
--plot_loss \
--bf16 \
--flash_attn fa2 \
--gradient_checkpointing \
--ddp_timeout 180000000 \
--deepspeed deepspeed/ds_zero3_offload_optimizer_param.json

Expected behavior

期望返回的结果形如下: image

Others

No response