Open dingdongwang opened 9 months ago
Should I update the eval_mdl_path to the path of my fine-tuned checkpoint to run batch inference, while keeping the base_model: str = "../../pretrained_mdls/vicuna_ltuas/" unchanged? Is my understanding correct?
Besides, if the following code means to merge the components of the LLM with the fine-tuned checkpoint (in another words, does the following code segment integrate the parts of the LLM with the fine-tuned parameters together as a new model?)
This is correct. str = "../../pretrained_mdls/vicuna_ltuas/"
is to load LLM and pretrained audio encoder, which actually happens here:
The state_dict = torch.load(eval_mdl_path, map_location='cpu')
is to further load finetuned audio encoder, linear projection layer, and lora adapters.
print('unexpect', unexpect)
This should not print anything.
If you do not fully understand how HF loading working, I recommend to keep this loading scheme. HF loading is not just loading a checkpoint.
-Yuan
Got it, thank you so much!
Hi, I have another question about the model related configuration settings during batch inference after model fine tuning.
In the inference_batch.py script for LTU-AS provided below:
Should I update the eval_mdl_path to the path of my fine-tuned checkpoint to run batch inference, while keeping the base_model: str = "../../pretrained_mdls/vicuna_ltuas/" unchanged? Is my understanding correct?
Besides, if the following code means to merge the components of the LLM with the fine-tuned checkpoint (in another words, does the following code segment integrate the parts of the LLM with the fine-tuned parameters together as a new model?)
Thank you so much for your patience and detailed responses all the time!