Open 0wwafa opened 4 months ago
thats because you are trying to merge models with different underlying base models. wizardlm2 is not a finetune of mistral v0.3, it's a finetune of v0.1, so its not surprising that it doesn't work since they are different base models.
see #324
well.. there should be a flag or soething to make that possible...
Considering I have metered internet and not so great resources, I followed your guind and the notebook. I used this yaml:
after more than an hour on colab I uploaded the merged model to huggingface. https://huggingface.co/ZeroWw/ZeroWw-mwiz-7B-slerp
Then I quantized it and downloaded (painfully, from where I am).
The result:
Meh.