Open 20241123 opened 2 months ago
I am trying to work with Jiutian 13.9b MoE model.But getting error in model compilation step.
Steps to reproduce the behavior:
1. pip install --pre -U -f https://mlc.ai/wheels mlc-chat-nightly-cu122 mlc-ai-nightly-cu122 git lfs install git clone https://huggingface.co/clemenschen/Jiutian-139MoE-chat-4bit mlc_llm convert_weight Jiutian-139MoE-chat --quantization q4f16_1 -o Jiutian-139MoE-chat_mlc
Results in:
raise ValueError(f"Unknown model type: {model_type}. Available ones: {list(MODELS.keys())}") ValueError: Unknown model type: jiutian. Available ones: ['llama', 'mistral', 'gemma', 'gemma2', 'gpt2', 'mixtral', 'gpt_neox', 'gpt_bigcode', 'phi-msft', 'phi', 'phi3', 'phi3_v', 'qwen', 'qwen2', 'qwen2_moe', 'stablelm', 'baichuan', 'internlm', 'internlm2', 'rwkv5', 'orion', 'llava', 'rwkv6', 'chatglm', 'eagle', 'bert', 'medusa', 'starcoder2', 'cohere', 'minicpm']
Should convert the weights. It works fine with qwen and qwen2.
Operating system: Ubuntu Device: v100 How you installed MLC-LLM: pip AND source How you installed TVM-Unity: pip Python version: 3.10 GPU driver version (if applicable): 5.7
Hi @20241123 thanks for the question. This model is in a new architecture that MLC does not yet support. Contributions are welcome :-)
🐛 Bug
I am trying to work with Jiutian 13.9b MoE model.But getting error in model compilation step.
To Reproduce
Steps to reproduce the behavior:
1. pip install --pre -U -f https://mlc.ai/wheels mlc-chat-nightly-cu122 mlc-ai-nightly-cu122 git lfs install git clone https://huggingface.co/clemenschen/Jiutian-139MoE-chat-4bit mlc_llm convert_weight Jiutian-139MoE-chat --quantization q4f16_1 -o Jiutian-139MoE-chat_mlc
Results in:
raise ValueError(f"Unknown model type: {model_type}. Available ones: {list(MODELS.keys())}") ValueError: Unknown model type: jiutian. Available ones: ['llama', 'mistral', 'gemma', 'gemma2', 'gpt2', 'mixtral', 'gpt_neox', 'gpt_bigcode', 'phi-msft', 'phi', 'phi3', 'phi3_v', 'qwen', 'qwen2', 'qwen2_moe', 'stablelm', 'baichuan', 'internlm', 'internlm2', 'rwkv5', 'orion', 'llava', 'rwkv6', 'chatglm', 'eagle', 'bert', 'medusa', 'starcoder2', 'cohere', 'minicpm']
Expected behavior
Should convert the weights. It works fine with qwen and qwen2.
Environment
Operating system: Ubuntu Device: v100 How you installed MLC-LLM: pip AND source How you installed TVM-Unity: pip Python version: 3.10 GPU driver version (if applicable): 5.7