Closed Ancientshi closed 1 year ago
Thanks for your interest in LMFlow! It could be caused by RAM optimized load. You may try to add --use_ram_optimized_load 0
in the run_finetune.sh
script and see if it works. Also, could you please provide your GPU type and memory size, so we may check for you if it is caused by lacking of GPU memory? Thanks 😄
Hi, thanks for your reply. Maybe I have already solved this issue by clearing the dataset cache. And I have another question: when we train the model using Reward Modeling or RAFT, can we use the LoRA method? or we should (1)train on the base model to make it aligned and then (2 LoRA training on an other domain?
Hi, thanks for your reply. Maybe I have already solved this issue by clearing the dataset cache. And I have another question: when we train the model using Reward Modeling or RAFT, can we use the LoRA method? or we should (1)train on the base model to make it aligned and then (2 LoRA training on an other domain?
Yes. You can use LoRA in RAFT.
Hi, thanks for your reply. Maybe I have already solved this issue by clearing the dataset cache. And I have another question: when we train the model using Reward Modeling or RAFT, can we use the LoRA method? or we should (1)train on the base model to make it aligned and then (2 LoRA training on an other domain?
Hello, I encounter the same question as you. I use dataset.cleanup_cache_files() to clear the dataset cache, but I still have the question above, may I ask how do you clear your dataset cache or how do you solve the problem, many thanks.
Hi, thanks for your reply. Maybe I have already solved this issue by clearing the dataset cache. And I have another question: when we train the model using Reward Modeling or RAFT, can we use the LoRA method? or we should (1)train on the base model to make it aligned and then (2 LoRA training on an other domain?
Hello, I encounter the same question as you. I use dataset.cleanup_cache_files() to clear the dataset cache, but I still have the question above, may I ask how do you clear your dataset cache or how do you solve the problem, many thanks.
You may clean the huggingface data cache via rm -rf ~/.cache/huggingface/datasets
. For more details, please refer to this doc. Hope that answers your question. Thanks 😄
This issue has been marked as stale because it has not had recent activity. If you think this still needs to be addressed please feel free to reopen this issue. Thanks
Originally posted by @Ancientshi in https://github.com/OptimalScale/LMFlow/issues/114#issuecomment-1537455013