Closed NamburiSrinath closed 7 months ago
Hi @NamburiSrinath, thanks for raising an issue.
It's hard to tell from the information here - but this looks like an error relating to the mergekit library rather than anything to do with transformers.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
Have you solve the problem yet? I meet the same problem as yours.
Nope, unfortunately I didn't.
Feature request
Hi,
I am working on model merging using mergekit (https://huggingface.co/blog/mlabonne/merge-models) and tried to merge finetuned GPT2. But I was getting the following errors -
SLERP:
Error -
RuntimeError: Tensor ln_f.bias required but not present in model /hdd4/srinath2/Trading_Agent/model_merging/model2_training/checkpoint-3750
TIES
Error -
RuntimeError: Tensor score.weight required but not present in model openai-community/gpt2
Same error for DARENote: The moment I change the base model to one of the finetuned models, I am able to merge the finetuned GPT2 models. I am not sure if this is expected because the blogpost has a base model that's different from the other 2 models
Motivation
I finetuned GPT2 on 2 tasks (model1 and model2) and wanted to merge them to understand if the merged model is better than base model on these tasks
Your contribution
If there's any inconsistencies, I can contribute to blog editing/PR if needed!