Open winglian opened 1 month ago
Thank you @winglian for raising the issue! Looks like FLCE fails at torch autograd when we make embedding + peft layer trainable together. We are happy to take a look! Would be nice if you can provide complete reproducible script. Thanks!
Thank you @winglian for raising the issue! Looks like FLCE fails at torch autograd when we make embedding + peft layer trainable together. We are happy to take a look! Would be nice if you can provide complete reproducible script. Thanks!
Can we create a list of what is working and what is not, with the corresponding dates, and keep it updated? For example, the DPO training labels issue, the PEFT+ embedding training issue, etc. Also, indicate which ones are going to be fixed and which ones are not.
š Describe the bug
when trying to train both LoRA layers on the base model and also set modules_to_save on the lora config which makes the embeddings layers trainable (my assumption is it also applies a LoRA linear adapter to the nn.Embeddings), results in the error:
Reproduce
Versions
No response