Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
15.37k
stars
2.22k
forks
source link
Llama 3.2-11B-vision fully fine-tuned model file question #727
The model was not saved to the finetune/output folder I specified, and moreover, the model weight files appear as follows, preventing me from performing inference.
ls
__0_0.distcp __1_0.distcp __2_0.distcp __3_0.distcp train_params.yaml
How can I save the weights of a fully fine-tuned model to a specified path, ensuring that the saved model weight file follows the standard transformers structure?
During the use of LoRA fine-tuning, everything was normal, but the following issue arose during full-scale fine-tuning.
I use the following script for full fine-tuning :
The model was not saved to the
finetune/output
folder I specified, and moreover, the model weight files appear as follows, preventing me from performing inference.How can I save the weights of a fully fine-tuned model to a specified path, ensuring that the saved model weight file follows the standard transformers structure?