Closed Ma1oneZhang closed 2 months ago
You need to first use the tools provided by InternLM to convert the internlm2_5-7b-chat-1m model into the LLaMA format. After that, you can convert it to a GGUF file. When using this method to download the model, please note that the model configuration needs to be in the LLaMA format, and the tokenizer should use the InternLM version.
As the tutorial here mentions, we have uploaded the model configuration, GGUF file, and tokenizer to a Hugging Face repository. We recommend downloading the model directly from this link to avoid the cumbersome process of merging the related files.
You need to first use the tools provided by InternLM to convert the internlm2_5-7b-chat-1m model into the LLaMA format. After that, you can convert it to a GGUF file. When using this method to download the model, please note that the model configuration needs to be in the LLaMA format, and the tokenizer should use the InternLM version.
As the tutorial here mentions, we have uploaded the model configuration, GGUF file, and tokenizer to a Hugging Face repository. We recommend downloading the model directly from this link to avoid the cumbersome process of merging the related files.
Do you plan to support loading the files from huggingface like AutoConfig
in transformers?
It's likely coming at some point, but not anytime soon.
Thank you!
How can i run internlm2_5-7b-chat-1m in ktransformers?
I have downloaded the internlm2_5-7b-chat-1m model file and converted it to a gguf file. But ktransformer still reports an error like that: