(venv) C:\LLaMA-Factory>python initilize_qpissa.py --base_model_dir NousResearch/Meta-Llama-3-8B-Instruct --output_path Meta-Llama-3-8B-Instruct-pissa-4bit-r128-iter4 --iter 4
bin C:\LLaMA-Factory\venv\lib\site-packages\bitsandbytes\libbitsandbytes_cuda121.dll
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [01:12<00:00, 18.15s/it]
Special tokens have been added in the vocabulary, make sure the associated word embeddings are fine-tuned or trained.
0%| | 1/739 [00:03<41:03, 3.34s/it]
Traceback (most recent call last):
File "C:\LLaMA-Factory\initilize_qpissa.py", line 86, in <module>
base_layer_in_4bits, base_layer, lora_A, lora_B = pissa_quant(value, args.rank, args.iter)
File "C:\LLaMA-Factory\venv\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "C:\LLaMA-Factory\initilize_qpissa.py", line 57, in pissa_quant
weight_nf4, weight_dequantized = quantize_and_dequantized(res)
File "C:\LLaMA-Factory\initilize_qpissa.py", line 44, in quantize_and_dequantized
weight_dequantized = bnb.functional.dequantize_4bit(
File "C:\LLaMA-Factory\venv\lib\site-packages\bitsandbytes\functional.py", line 1018, in dequantize_4bit
assert absmax is not None and out is not None
AssertionError
Can the quantized conversion feature of PiSSA be implemented in LLaMA-Factory for QLoRA?
Reminder
System Info
llamafactory
version: 0.8.3.dev0Reproduction
--finetuning_type lora
+--quantization_bit 4
+--pissa_init True
Expected behavior
QLoRA and PISSA are not directly supported, scripts/pissa_init.py is not a available script, instead we should use initilize_qpissa.py?
Can the quantized conversion feature of PiSSA be implemented in LLaMA-Factory for QLoRA?
Others
No response