Closed xyangyan closed 2 months ago
Depending on what you did to the architecture you need to change https://github.com/ggerganov/llama.cpp/blob/822b6322dea704110797a5671fc80ae39ee6ac97/convert_hf_to_gguf.py#L1490-L1635
Adding "LlamaPrunedForCausalLM" to the register in the first line could be enough. If it's not then you need to implement a separate class with your changes that could inherit after LlamaModel
What happened?
python convert_hf_to_gguf.py /llama3.1_pruned_0.6_4-29
We executed the conversion command, but encountered the above error. Here, 'llama3.1_pruned_0.6_4-29' is the version we used after structured pruning. I would like to ask how to operate to convert it to the gguf format, and which file's code needs to be adapted?Name and Version
version: 3735 (df4b7945) built with cc (Ubuntu 9.4.0-1ubuntu1~20.04.2) 9.4.0 for x86_64-linux-gnu
What operating system are you seeing the problem on?
Linux
Relevant log output
No response