lxe / llama-tune

LLaMa Tuning with Stanford Alpaca Dataset using Deepspeed and Transformers
51 stars 7 forks source link

llama-7b-hf Tuning with Stanford Alpaca Dataset using Deepspeed and Transformers

This is my first go at ML tuning, so this is probably very wrong. This should work on a single 3090 GPU A100 and takes 3 hours to train 250 setps on a subset of 1000 samples. Full 50k~ dataset should take ~19 hours. There's a lot of knobs to turn that I don't understand yet.

I'm currently running this, so I don't know if it even works

References:

Prereqs

conda install -y cuda pytorch-cuda=11.7 -c pytorch -c nvidia

Run tuning

git clone https://github.com/tatsu-lab/stanford_alpaca repositories/stanford_alpaca
pip install -r requirements.txt
python download-model.py decapoda-research/llama-7b-hf
deepspeed tune.py

Comparing the model tuned on 250-step 1000 training samples VS vanilla llama-7b:

image

Vanilla llama-7b:

image