Closed spiritfog closed 5 months ago
Seems the error indicates it cannot find nvcc in your env. Make sure you install cuda and nvcc is in your PATH
Seems the error indicates it cannot find nvcc in your env. Make sure you install cuda and nvcc is in your PATH
3Q, I would have another try now!
🐛 Bug TVM ERROR when convert_weight
llava model convert_weight failed, especially when quantization by tvm
To Reproduce
Steps to reproduce the behavior:
Expected behavior
successful quantize the llava-1.5-7b model and convert it to MLC form,
Environment
conda
, source): yespip
, source): installed mlc_aipython -c "import tvm; print('\n'.join(f'{k}: {v}' for k, v in tvm.support.libinfo().items()))"
, applicable if you compile models): f5f048bbd71513f087799f987019e3931f68a6d9Additional context
1. I think that error may rised by tvm, so verified it by tvm verify All verification is passed, except for the detection of the Vulkan device. When I input
python -c "import tvm; print(tvm.vulkan().exist)"
, I got a error rather than True or False. the Error message:2. I think the error come from tvm, so try to convert model weight without quantization, when use
mlc_llm convert_weight /workspace/mlc_llm/download/llava-1.5-7b-hf/ --quantization q0f16 -o ./llava-1.5-7b-hf-MLC-q0f16
, there is no error