mit-han-lab / llm-awq

[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
MIT License
2.08k stars 150 forks source link

Error while generating real quantized weights for VILA #160

Open ocg2347 opened 3 months ago

ocg2347 commented 3 months ago

I can successfully run vila-7b but when i want to generate awq weights using the "vila-7b-w4-g128-v2.pt" from "https://huggingface.co/Efficient-Large-Model/VILA-7b-4bit-awq/tree/main" I got the error below. Anyone facing this? or anyone who managed to get inference with vila-7b-awq?

note to developers: In "https://github.com/mit-han-lab/llm-awq/blob/main/scripts/vila_example.sh", it is stated that search results are shared, but it does not seem under "https://huggingface.co/datasets/mit-han-lab/awq-model-zoo/tree/main".

` root@e9118846cb22:/llm-awq# python -m awq.entry --model_path vila-7b --w_bit 4 --q_group_size 128 --load_awq VILA-7b-4bit-awq/vila-7b-w4-g128-v2.pt --q_backend real --dump_quant quant_cache/llama-2-7b-chat-w4-g128-awq.pt Quantization config: {'zero_point': True, 'q_group_size': 128}