Open erlakshmi123 opened 1 year ago
Hi! Have you been getting any errors/warning alongside this?
I have not see any errors.
Hi ! @erlakshmi123 Try to use a small value for the batch_size (8, 16, 32) and epoch (8, 10) if u have not enough dataset. It should work.
I have tried multiple attempts with this I always get the adapter_model.bin as 400 bytes.. looks like its not training or saving the model. train data is small ~2000 json lines in dolly prompt, response format.
python src/finetune.py \ --base_model 'mosaicml/mpt-7b-instruct' \ --data_path 'dataset/train_data.json' \ --output_dir './lora-mpt' \ --batch_size 256 \ --micro_batch_size 4 \ --num_epochs 100 \ --learning_rate 3e-5 \ --cutoff_len 1024 \ --val_set_size 200 \ --lora_r 4 \ --lora_alpha 8 \ --lora_dropout 0.05 \ --lora_target_modules '[Wqkv]' \ --train_on_inputs False \ --group_by_length False \ --use_gradient_checkpointing True \ --load_in_8bit False \ --needs_prompt_generation False