Open tomgm777 opened 3 weeks ago
INFO:hf-to-gguf:Loading model: qwen
INFO:gguf.gguf_writer:gguf: This GGUF file is for Little Endian only
INFO:hf-to-gguf:Set model parameters
INFO:hf-to-gguf:Set model tokenizer
INFO:gguf.vocab:Adding 151387 merge(s).
INFO:gguf.vocab:Setting special token type bos to 151643
INFO:gguf.vocab:Setting special token type eos to 151643
INFO:gguf.vocab:Setting special token type unk to 151643
INFO:hf-to-gguf:Exporting model...
INFO:hf-to-gguf:gguf: loading model weight map from 'model.safetensors.index.json'
INFO:hf-to-gguf:gguf: loading model part 'model-00001-of-00003.safetensors'
INFO:hf-to-gguf:blk.0.attn_qkv.bias, torch.float16 --> F32, shape = {12288}
Traceback (most recent call last):
File "D:\AI\llama.cpp-master\convert-hf-to-gguf.py", line 3263, in
What happened?
When I tried to quantize using the following command, I got the following error. Do you know the cause?
py convert-hf-to-gguf.py --outtype f16 F:/models/Llama-3-Lumimaid-70B-v0.1-alt/
Name and Version
today (7/2) git clone version
What operating system are you seeing the problem on?
Windows
Relevant log output
No response