[X] I have read the README and searched the existing issues.
System Info
I'm using the latest llamafactory version.
Reproduction
Hi, I'm trying to use qlora+fsdp for the llama3.1 70B model with 8xA100(640G). In my case, the context length is large(32k), I'm using the fsdp and qlora setting(quantization_bit=4) with fa2, but still get OOM after some iterations. Is there any other way to save the memory further?
Reminder
System Info
I'm using the latest llamafactory version.
Reproduction
Hi, I'm trying to use qlora+fsdp for the llama3.1 70B model with 8xA100(640G). In my case, the context length is large(32k), I'm using the fsdp and qlora setting(quantization_bit=4) with fa2, but still get OOM after some iterations. Is there any other way to save the memory further?
Expected behavior
No response
Others
No response