Closed wenjingk-xilinx closed 7 months ago
Hi, the issue with data type conflicts has been resolved. You can simply use the default parameters in qalora.py for fine-tuning.
Hi @xxw11 , thank you for your update. I see you've changed the data format into fp32. The training becomes much slower with fp32 than fp16. Have you compared the accuracy with these two different data types?
Yes, you're right, the current code will be slower. In my experiments, using FP32 results in higher accuracy, especially on Llama1.
Hi, I tried to run the code with this script:
F.linear(input, self.weight, self.bias) raises an error, input.dtype=float32, self.weight.dtype=float16.
Then I add '--fp16' flag in the running script, the bug becomes: