VinAIResearch / PhoGPT

PhoGPT: Generative Pre-training for Vietnamese (2023)
Apache License 2.0
739 stars 67 forks source link

Error message: "llama.cpp error: 'done_getting_tensors: wrong number of tensors; expected 388, got 195'" #23

Closed jackNhat closed 5 months ago

jackNhat commented 6 months ago

I converted the model into the llama.cpp's own format (.GGUF). And this model could run successful in colab like snapshot.

PhoGPT-error

But when i loaded this model in LLM Studio or Jan, i got error message. Anyone met this problem? How to resolve this?

LLMStudio-error

Many thanks!

xtfocus commented 6 months ago

Likely a duplicated of my opening issue #22

datquocnguyen commented 6 months ago

See: https://github.com/ggerganov/llama.cpp/pull/5638

datquocnguyen commented 6 months ago

PR was merged. It should work fine now.

churongcon commented 5 months ago

image @datquocnguyen Not working

001123 commented 5 months ago
image

Same issue

datquocnguyen commented 5 months ago

@001123 @churongcon Please could you try installing/compiling llama.cpp from this specific commit and then redo the conversion step (to create the .gguf file):

https://github.com/jordankanter/llama.cpp/tree/87a41f53ae3a01dde4c198df72cfb99ba2c9f586

git clone https://github.com/jordankanter/llama.cpp
git checkout 87a41f53ae3a01dde4c198df72cfb99ba2c9f586
datquocnguyen commented 5 months ago

@001123 @churongcon, please disregard my previous comment. I have found the reason. "llama.cpp" has been modified, resulting in a converted version of phogpt-4b being associated with 387 tensors, instead of the previous 388. To load it with the latest version of Studio, you can simply reconvert phogpt-4b using the latest "llama.cpp".

datquocnguyen commented 5 months ago

@001123 @churongcon please try https://huggingface.co/vinai/PhoGPT-4B-Chat-gguf