tloen / alpaca-lora

Instruct-tune LLaMA on consumer hardware
Apache License 2.0
18.55k stars 2.21k forks source link

Low memory crash while loading shards #117

Open francis2tm opened 1 year ago

francis2tm commented 1 year ago

Hello, Clearly my PC is not as juicy as yours. I'm attempting to load the model with LlamaForCausalLM.from_pretrained() but the script crashes since I don't have enough RAM (I only have 16GB). Is there anyway to load with only 16GB?

Thanks in advance

CarlDegio commented 1 year ago

Seems the lora train need 18G: https://twitter.com/nash_su/status/1637423768665718784 And I would like to know the generate.py also need 18G vRAM?