Closed bkhanal-11 closed 1 week ago
Hi @bkhanal-11 for the single gpu use case this might work, but for the advances use cases with FSDP you'll need to different wrapping policy and there might be other things to consider and adapt.
That makes sense, thanks.
Thank you for your continuous support to the LLM open-source community.
I was wondering if we use
AutoModel
,AutoConfig
andAutoTokenizer
instead ofLlamaForCausalLM
,LlamaConfig
andLlamaTokenizer
for general LLM fine-tuning, will the fine-tuning pipeline/recipe fail?