SJTU-IPADS / PowerInfer

High-speed Large Language Model Serving on PCs with Consumer-grade GPUs
MIT License
7.98k stars 415 forks source link

How to convert ProSparse-LLaMA-2-13B model to .gguf? #201

Open Graysonicc opened 5 months ago

Graysonicc commented 5 months ago

Prerequisites

Before submitting your question, please ensure the following:

Question Details

Please provide a clear and concise description of your question. If applicable, include steps to reproduce the issue or behaviors you've observed.

Additional Context

Please provide any additional information that may be relevant to your question, such as specific system configurations, environment details, or any other context that could be helpful in addressing your inquiry.

Graysonicc commented 5 months ago

python convert.py --outfile ./llama_convert/prosparse-llama-2-13b.powerinfer.gguf ./prosparse-llama-2-13b ./prosparse-llama-2-13b-predictor Model architecture True is not supported by this convert.py. Trying with convert-hf-to-powerinfer-gguf.py... Loading model: prosparse-llama-2-13b Traceback (most recent call last): File "/root/autodl-tmp/powerinfer/PowerInfer/convert-hf-to-powerinfer-gguf.py", line 609, in model_class = Model.from_model_architecture(hparams["architectures"][0]) File "/root/autodl-tmp/powerinfer/PowerInfer/convert-hf-to-powerinfer-gguf.py", line 189, in from_model_architecture raise NotImplementedError(f'Architecture "{model_architecture}" not supported!') NotImplementedError: Architecture "SparseLlamaForCausalLM" not supported!

Graysonicc commented 5 months ago

![Uploading @YSU)%~3F0F[]DI%4LW_COM.png…]()

baibizhe commented 5 months ago

Hello. Have you solve this bug ?