I use 1 node 4 V100(32G) to sft on llama2-7b, script: minillm/scripts/llama2/sft/sft_7B.sh, but got out of memory error. Should I using a more larger GPU? or I have something wrong? (because https://github.com/microsoft/LMOps/issues/91 is trained on V100)
I use 1 node 4 V100(32G) to sft on llama2-7b, script:
minillm/scripts/llama2/sft/sft_7B.sh
, but got out of memory error. Should I using a more larger GPU? or I have something wrong? (because https://github.com/microsoft/LMOps/issues/91 is trained on V100)