Open MoonRide303 opened 1 month ago
I'm also having this issue with the v3 tokenizer
@MoonRide303 I solved it by using consolidated.safetensors
It's in the mistral repo.
Make sure to delete the other safetensors first.
Thanks @arch-btw that seemed to do the trick
I just saw the same error when trying to convert a GPT2LMHeadModel (striki-ai/william-shakespeare-poetry).
@MoonRide303 I solved it by using
consolidated.safetensors
It's in the mistral repo. Make sure to delete the other safetensors first.
Yeah, having both consolidated and separated model files in the same folder causes the problem - I've moved consolidated.safetensors to different folder, and conversion started working, then, too.
But it looks like it should be corrected on llama.cpp side, too - model.safetensors.index.json
file (which defines which model files should be used) is being ignored.
I was trying to convert recently released Mistral v0.3 models (Mistral-7B-v0.3, Mistral-7B-Instruct-v0.3) using
python convert-hf-to-gguf.py --outtype f16 ..\Mistral-7B-Instruct-v0.3\ --outfile Mistral-7B-Instruct-v0.3-F16.gguf
command from llama.cpp b2972, but in both cases I end up with:Mistral v0.2 conversion works fine. Running Q6_K quant (provided by @MaziyarPanahi) from Mistral-7B-Instruct-v0.3-GGUF repo works fine, too.