Closed wum67 closed 10 months ago
Hi,
Please use the following command to train LoRA:
CUDA_VISIBLE_DEVICES=0 python finetune.py --base_model 'yahma/llama-7b-hf' --data_path 'math_10K.json' --output_dir './trained_models/llama-7b-lora-math/' --batch_size 16 --micro_batch_size 4 --num_epochs 3 --learning_rate 3e-4 --cutoff_len 256 --val_set_size 120 --eval_step 80 --save_step 80 --adapter_name lora --target_modules '["q_proj", "k_proj", "v_proj", "up_proj", "down_proj"]' --lora_r 32 --lora_alpha 64
Here's how I trained the model:
Here's how I evaluated the model on SWAMP:
I got a 0% accuracy and a lot of times the model is over generating the predictions. For example:
Is there anything I'm doing wrong?