Open dmsweetser opened 5 months ago
Same error with https://huggingface.co/pjh64/Phi-3-mini-128K-Instruct.gguf
That's because the .gguf files do seem to be corrupted or bad. The gguf files from here work flawless https://huggingface.co/PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed
It looks like they are still working on the 128k variant in llama.cpp:
One other note - I am seeing that I can load https://huggingface.co/PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed, but there are irregularities when I use a larger prompt. I suspect this is related to the llama.cpp remaining work.
Describe the bug
I'm attempting to load https://huggingface.co/QuantFactory/Phi-3-mini-128k-instruct-GGUF with default options and I receive this error:
Traceback (most recent call last):
File "C:\Files\text-generation-web-ui\modules\ui_model_menu.py", line 247, in load_model_wrapper
shared.model, shared.tokenizer = load_model(selected_model, loader)
File "C:\Files\text-generation-web-ui\modules\models.py", line 94, in load_model
output = load_func_maploader
File "C:\Files\text-generation-web-ui\modules\models.py", line 271, in llamacpp_loader
model, tokenizer = LlamaCppModel.from_pretrained(model_file)
File "C:\Files\text-generation-web-ui\modules\llamacpp_model.py", line 77, in from_pretrained
result.model = Llama(**params)
File "C:\Files\text-generation-web-ui\installer_files\env\Lib\site-packages\llama_cpp\llama.py", line 323, in init
self._model = _LlamaModel(
File "C:\Files\text-generation-web-ui\installer_files\env\Lib\site-packages\llama_cpp_internals.py", line 55, in init
raise ValueError(f"Failed to load model from file: {path_model}") ValueError: Failed to load model from file: models\Phi-3-mini-128k-instruct.Q4_K_M.gguf
Is there an existing issue for this?
Reproduction
Download and load the model:
https://huggingface.co/QuantFactory/Phi-3-mini-128k-instruct-GGUF/resolve/main/Phi-3-mini-128k-instruct.Q4_K_M.gguf?download=true
Screenshot
No response
Logs
System Info