Closed greatzane closed 2 days ago
hi, the logging here is not a warning or issue from green-bit-llm, instead, it is a warning from the huggingface library. This is because there is no bias for down/gate/up/o_proj
layers in transformers block for Qwen2. https://github.com/huggingface/transformers/blob/9fe3f585bb4ea29f209dc705d269fbe292e1128f/src/transformers/models/qwen2/modeling_qwen2.py#L174. This explains why thebias
term in Qwen2-0.5B-Instruct-GPTQ-Int4 is not used for transformers AutoModelForCausalLM
.
If you wanna try finetuning GPTQ model, please follow the below command:
CUDA_VISIBLE_DEVICES=0 python -m green_bit_llm.sft.peft_lora --model astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit --dataset tatsu-lab/alpaca --lr-fp 1e-6
CUDA_VISIBLE_DEVICES=0 python -m green_bit_llm.sft.finetune --model astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit --dataset tatsu-lab/alpaca --tune-qweight-only --batch-size 1
Hello, I am encountering this message when finetuning with Qwen2 GPTQ Int4 series models:
And the finetuned model cannot be use for inference.
The command line is:
python -m green_bit_llm.sft.finetune --model ../Qwen2-0.5B-Instruct-GPTQ-Int4/ --dataset tatsu-lab/alpaca --batch-size 1 --tune-qweight-only
Can you explain how to fix this? Thanks!