Open goodnight654 opened 5 months ago
I have never been able to correctly quantify awq for llava-llama3 in the official format of llava。 Can anyone help me?
autoawq appears to support quantization for llava models, have you tried it?
autoawq
We are developing a quantized version of the VL model in lmdeploy, but it won't be released until next week.
lmdeploy
I have never been able to correctly quantify awq for llava-llama3 in the official format of llava。 Can anyone help me?